Shaoting Zhang , Yishan Du , Wenji Wang , Xianying He , Fangfang Cui , Liang Zhao , Bei Wang , Zhiqiang Hu , Ziqiang Wang , Qing Xia , Tian Shen , Jie Zhao
{"title":"ECGFM: A foundation model for ECG analysis trained on a multi-center million-ECG dataset","authors":"Shaoting Zhang , Yishan Du , Wenji Wang , Xianying He , Fangfang Cui , Liang Zhao , Bei Wang , Zhiqiang Hu , Ziqiang Wang , Qing Xia , Tian Shen , Jie Zhao","doi":"10.1016/j.inffus.2025.103363","DOIUrl":null,"url":null,"abstract":"<div><div>The electrocardiogram (ECG) is widely used for diagnosing heart conditions due to its cost-effectiveness, non-invasiveness, and accessibility. Between 2014 and 2017, the First Affiliated Hospital of Zhengzhou University collected over a million clinical ECGs from diverse primary hospitals, each accompanied by initial diagnostic results. Effectively utilizing this vast dataset with potential label inconsistencies is a key challenge. In this study, we introduce ECGFM, a foundation model pre-trained on over a million clinical ECGs to achieve deep ECG comprehension. ECGFM comprises a convolutional encoder, a transformer decoder, and task-specific heads, pre-trained through three complementary sub-tasks: (i) contrastive predictive learning for unsupervised representation learning, (ii) normal/abnormal classification, and (iii) diagnostic text generation. Given potential label unreliability, active learning is integrated with the classification task to select key data for re-annotation, enhancing supervision quality. To enable ECGFM’s adaptability to downstream tasks with any-lead inputs, a transferred convolutional encoder is trained to align feature distributions. ECGFM’s effectiveness is evaluated using diverse public datasets, including PTB-XL, Georgia, CPSC, CinC 2020, MITDB, and the “Hefei High-tech Cup” dataset. Fine-tuning ECGFM (training only a task-specific head) delivers strong performance across datasets, nearing fully supervised methods. Additionally, in a one-month online test with 7951 recordings, ECGFM achieved a recall of 0.9335, a precision of 0.9571, and an F1-score of 0.9451, underscoring its robustness and potential for real-world applications.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"124 ","pages":"Article 103363"},"PeriodicalIF":14.7000,"publicationDate":"2025-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525004361","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The electrocardiogram (ECG) is widely used for diagnosing heart conditions due to its cost-effectiveness, non-invasiveness, and accessibility. Between 2014 and 2017, the First Affiliated Hospital of Zhengzhou University collected over a million clinical ECGs from diverse primary hospitals, each accompanied by initial diagnostic results. Effectively utilizing this vast dataset with potential label inconsistencies is a key challenge. In this study, we introduce ECGFM, a foundation model pre-trained on over a million clinical ECGs to achieve deep ECG comprehension. ECGFM comprises a convolutional encoder, a transformer decoder, and task-specific heads, pre-trained through three complementary sub-tasks: (i) contrastive predictive learning for unsupervised representation learning, (ii) normal/abnormal classification, and (iii) diagnostic text generation. Given potential label unreliability, active learning is integrated with the classification task to select key data for re-annotation, enhancing supervision quality. To enable ECGFM’s adaptability to downstream tasks with any-lead inputs, a transferred convolutional encoder is trained to align feature distributions. ECGFM’s effectiveness is evaluated using diverse public datasets, including PTB-XL, Georgia, CPSC, CinC 2020, MITDB, and the “Hefei High-tech Cup” dataset. Fine-tuning ECGFM (training only a task-specific head) delivers strong performance across datasets, nearing fully supervised methods. Additionally, in a one-month online test with 7951 recordings, ECGFM achieved a recall of 0.9335, a precision of 0.9571, and an F1-score of 0.9451, underscoring its robustness and potential for real-world applications.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.