Tiarna Lee, Esther Puyol-Antón, Bram Ruijsink, Sebastien Roujol, Theodore Barfoot, Shaheim Ogbomo-Harmitt, Miaojing Shi, Andrew King
{"title":"基于人工智能的心脏磁共振分割中种族偏差原因的研究。","authors":"Tiarna Lee, Esther Puyol-Antón, Bram Ruijsink, Sebastien Roujol, Theodore Barfoot, Shaheim Ogbomo-Harmitt, Miaojing Shi, Andrew King","doi":"10.1093/ehjdh/ztaf008","DOIUrl":null,"url":null,"abstract":"<p><strong>Aims: </strong>Artificial intelligence (AI) methods are being used increasingly for the automated segmentation of cine cardiac magnetic resonance (CMR) imaging. However, these methods have been shown to be subject to race bias; i.e. they exhibit different levels of performance for different races depending on the (im)balance of the data used to train the AI model. In this paper, we investigate the source of this bias, seeking to understand its root cause(s).</p><p><strong>Methods and results: </strong>We trained AI models to perform race classification on cine CMR images and/or segmentations from White and Black subjects from the UK Biobank and found that the classification accuracy for images was higher than for segmentations. Interpretability methods showed that the models were primarily looking at non-heart regions. Cropping images tightly around the heart caused classification accuracy to drop to almost chance level. Visualizing the latent space of AI segmentation models showed that race information was encoded in the models. Training segmentation models using cropped images reduced but did not eliminate the bias. A number of possible confounders for the bias in segmentation model performance were identified for Black subjects but none for White subjects.</p><p><strong>Conclusion: </strong>Distributional differences between annotated CMR data of White and Black races, which can lead to bias in trained AI segmentation models, are predominantly image-based, not segmentation-based. Most of the differences occur in areas outside the heart, such as subcutaneous fat. These findings will be important for researchers investigating performance of AI models on different races.</p>","PeriodicalId":72965,"journal":{"name":"European heart journal. Digital health","volume":"6 3","pages":"350-358"},"PeriodicalIF":4.4000,"publicationDate":"2025-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12088717/pdf/","citationCount":"0","resultStr":"{\"title\":\"An investigation into the causes of race bias in artificial intelligence-based cine cardiac magnetic resonance segmentation.\",\"authors\":\"Tiarna Lee, Esther Puyol-Antón, Bram Ruijsink, Sebastien Roujol, Theodore Barfoot, Shaheim Ogbomo-Harmitt, Miaojing Shi, Andrew King\",\"doi\":\"10.1093/ehjdh/ztaf008\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Aims: </strong>Artificial intelligence (AI) methods are being used increasingly for the automated segmentation of cine cardiac magnetic resonance (CMR) imaging. However, these methods have been shown to be subject to race bias; i.e. they exhibit different levels of performance for different races depending on the (im)balance of the data used to train the AI model. In this paper, we investigate the source of this bias, seeking to understand its root cause(s).</p><p><strong>Methods and results: </strong>We trained AI models to perform race classification on cine CMR images and/or segmentations from White and Black subjects from the UK Biobank and found that the classification accuracy for images was higher than for segmentations. Interpretability methods showed that the models were primarily looking at non-heart regions. Cropping images tightly around the heart caused classification accuracy to drop to almost chance level. Visualizing the latent space of AI segmentation models showed that race information was encoded in the models. Training segmentation models using cropped images reduced but did not eliminate the bias. A number of possible confounders for the bias in segmentation model performance were identified for Black subjects but none for White subjects.</p><p><strong>Conclusion: </strong>Distributional differences between annotated CMR data of White and Black races, which can lead to bias in trained AI segmentation models, are predominantly image-based, not segmentation-based. Most of the differences occur in areas outside the heart, such as subcutaneous fat. These findings will be important for researchers investigating performance of AI models on different races.</p>\",\"PeriodicalId\":72965,\"journal\":{\"name\":\"European heart journal. Digital health\",\"volume\":\"6 3\",\"pages\":\"350-358\"},\"PeriodicalIF\":4.4000,\"publicationDate\":\"2025-02-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12088717/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"European heart journal. Digital health\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1093/ehjdh/ztaf008\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/5/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q1\",\"JCRName\":\"CARDIAC & CARDIOVASCULAR SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"European heart journal. Digital health","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1093/ehjdh/ztaf008","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/5/1 0:00:00","PubModel":"eCollection","JCR":"Q1","JCRName":"CARDIAC & CARDIOVASCULAR SYSTEMS","Score":null,"Total":0}
An investigation into the causes of race bias in artificial intelligence-based cine cardiac magnetic resonance segmentation.
Aims: Artificial intelligence (AI) methods are being used increasingly for the automated segmentation of cine cardiac magnetic resonance (CMR) imaging. However, these methods have been shown to be subject to race bias; i.e. they exhibit different levels of performance for different races depending on the (im)balance of the data used to train the AI model. In this paper, we investigate the source of this bias, seeking to understand its root cause(s).
Methods and results: We trained AI models to perform race classification on cine CMR images and/or segmentations from White and Black subjects from the UK Biobank and found that the classification accuracy for images was higher than for segmentations. Interpretability methods showed that the models were primarily looking at non-heart regions. Cropping images tightly around the heart caused classification accuracy to drop to almost chance level. Visualizing the latent space of AI segmentation models showed that race information was encoded in the models. Training segmentation models using cropped images reduced but did not eliminate the bias. A number of possible confounders for the bias in segmentation model performance were identified for Black subjects but none for White subjects.
Conclusion: Distributional differences between annotated CMR data of White and Black races, which can lead to bias in trained AI segmentation models, are predominantly image-based, not segmentation-based. Most of the differences occur in areas outside the heart, such as subcutaneous fat. These findings will be important for researchers investigating performance of AI models on different races.