Snapshot artificial intelligence—determination of ejection fraction from a single frame still image: a multi-institutional, retrospective model development and validation study
Jeffrey G Malins PhD , D M Anisuzzaman PhD , John I Jackson PhD , Eunjung Lee PhD , Jwan A Naser MBBS , Behrouz Rostami PhD , Jared G Bird MD , Dan Spiegelstein MD , Talia Amar MSc , Prof Jae K Oh MD , Prof Patricia A Pellikka MD , Jeremy J Thaden MD , Prof Francisco Lopez-Jimenez MD MSc , Prof Sorin V Pislaru MD PhD , Prof Paul A Friedman MD , Prof Garvan C Kane MD PhD , Zachi I Attia PhD
{"title":"Snapshot artificial intelligence—determination of ejection fraction from a single frame still image: a multi-institutional, retrospective model development and validation study","authors":"Jeffrey G Malins PhD , D M Anisuzzaman PhD , John I Jackson PhD , Eunjung Lee PhD , Jwan A Naser MBBS , Behrouz Rostami PhD , Jared G Bird MD , Dan Spiegelstein MD , Talia Amar MSc , Prof Jae K Oh MD , Prof Patricia A Pellikka MD , Jeremy J Thaden MD , Prof Francisco Lopez-Jimenez MD MSc , Prof Sorin V Pislaru MD PhD , Prof Paul A Friedman MD , Prof Garvan C Kane MD PhD , Zachi I Attia PhD","doi":"10.1016/j.landig.2025.02.003","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>Artificial intelligence (AI) is poised to transform point-of-care practice by providing rapid snapshots of cardiac functioning. Although previous AI models have been developed to estimate left ventricular ejection fraction (LVEF), they have typically used video clips as input, which can be computationally intensive. In the current study, we aimed to develop an LVEF estimation model that takes in static frames as input.</div></div><div><h3>Methods</h3><div>Using retrospective transthoracic echocardiography (TTE) data from Mayo Clinic Rochester and Mayo Clinic Health System sites (training: n=19 627; interval validation: n=862), we developed a two-dimensional convolutional neural network model that provides an LVEF estimate associated with an input frame from an echocardiogram video. We then evaluated model performance for Mayo Clinic TTE data (Rochester, n=1890; Arizona, n=1695; Florida, n=1862), the EchoNet-Dynamic TTE dataset (n=10 015), a prospective cohort of patients from whom TTE and handheld cardiac ultrasound (HCU) were simultaneously collected (n=625), and a prospective cohort of patients from whom HCU clips were collected by expert sonographers and novice users (n=100, distributed across three external sites).</div></div><div><h3>Findings</h3><div>We observed consistently strong model performance when estimates from single frames were averaged across multiple video clips, even when only one frame was taken per video (for classifying LVEF ≤40% <em>vs</em> LVEF>40%, area under the receiver operating characteristic curve [AUC]>0·90 for all datasets except for HCU clips collected by novice users, for which AUC>0·85). We also observed that LVEF estimates differed slightly depending on the phase of the cardiac cycle when images were captured.</div></div><div><h3>Interpretation</h3><div>When aiming to rapidly deploy such models, single frames from multiple videos might be sufficient for LVEF classification. Furthermore, the observed sensitivity to the cardiac cycle offers some insights on model performance from an explainability perspective.</div></div><div><h3>Funding</h3><div>Internal institutional funds provided by the Department of Cardiovascular Medicine, Mayo Clinic, Rochester, MN, USA.</div></div>","PeriodicalId":48534,"journal":{"name":"Lancet Digital Health","volume":"7 4","pages":"Pages e255-e263"},"PeriodicalIF":23.8000,"publicationDate":"2025-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Lancet Digital Health","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2589750025000275","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MEDICAL INFORMATICS","Score":null,"Total":0}
引用次数: 0
Abstract
Background
Artificial intelligence (AI) is poised to transform point-of-care practice by providing rapid snapshots of cardiac functioning. Although previous AI models have been developed to estimate left ventricular ejection fraction (LVEF), they have typically used video clips as input, which can be computationally intensive. In the current study, we aimed to develop an LVEF estimation model that takes in static frames as input.
Methods
Using retrospective transthoracic echocardiography (TTE) data from Mayo Clinic Rochester and Mayo Clinic Health System sites (training: n=19 627; interval validation: n=862), we developed a two-dimensional convolutional neural network model that provides an LVEF estimate associated with an input frame from an echocardiogram video. We then evaluated model performance for Mayo Clinic TTE data (Rochester, n=1890; Arizona, n=1695; Florida, n=1862), the EchoNet-Dynamic TTE dataset (n=10 015), a prospective cohort of patients from whom TTE and handheld cardiac ultrasound (HCU) were simultaneously collected (n=625), and a prospective cohort of patients from whom HCU clips were collected by expert sonographers and novice users (n=100, distributed across three external sites).
Findings
We observed consistently strong model performance when estimates from single frames were averaged across multiple video clips, even when only one frame was taken per video (for classifying LVEF ≤40% vs LVEF>40%, area under the receiver operating characteristic curve [AUC]>0·90 for all datasets except for HCU clips collected by novice users, for which AUC>0·85). We also observed that LVEF estimates differed slightly depending on the phase of the cardiac cycle when images were captured.
Interpretation
When aiming to rapidly deploy such models, single frames from multiple videos might be sufficient for LVEF classification. Furthermore, the observed sensitivity to the cardiac cycle offers some insights on model performance from an explainability perspective.
Funding
Internal institutional funds provided by the Department of Cardiovascular Medicine, Mayo Clinic, Rochester, MN, USA.
期刊介绍:
The Lancet Digital Health publishes important, innovative, and practice-changing research on any topic connected with digital technology in clinical medicine, public health, and global health.
The journal’s open access content crosses subject boundaries, building bridges between health professionals and researchers.By bringing together the most important advances in this multidisciplinary field,The Lancet Digital Health is the most prominent publishing venue in digital health.
We publish a range of content types including Articles,Review, Comment, and Correspondence, contributing to promoting digital technologies in health practice worldwide.