减轻人工智能对少数族裔死亡率预测的偏见:一种迁移学习方法。

IF 3.3 3区 医学 Q2 MEDICAL INFORMATICS
Tianshu Gu, Wensen Pan, Jing Yu, Guang Ji, Xia Meng, Yongjun Wang, Minghui Li
{"title":"减轻人工智能对少数族裔死亡率预测的偏见:一种迁移学习方法。","authors":"Tianshu Gu, Wensen Pan, Jing Yu, Guang Ji, Xia Meng, Yongjun Wang, Minghui Li","doi":"10.1186/s12911-025-02862-7","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>The COVID-19 pandemic has highlighted the crucial role of artificial intelligence (AI) in predicting mortality and guiding healthcare decisions. However, AI models may perpetuate or exacerbate existing health disparities due to demographic biases, particularly affecting racial and ethnic minorities. The objective of this study is to investigate the demographic biases in AI models predicting COVID-19 mortality and to assess the effectiveness of transfer learning in improving model fairness across diverse demographic groups.</p><p><strong>Methods: </strong>This retrospective cohort study used a population-based dataset of COVID-19 cases from the Centers for Disease Control and Prevention (CDC), spanning the years 2020-2024. The study analyzed AI model performance across different racial and ethnic groups and employed transfer learning techniques to improve model fairness by adapting pre-trained models to the specific demographic and clinical characteristics of the population.</p><p><strong>Results: </strong>Decision Tree (DT) and Random Forest (RF) models consistently showed improvements in accuracy, precision, and ROC-AUC scores for Non-Hispanic Black, Hispanic/Latino, and Asian populations. The most significant precision improvement was observed in the DT model for Hispanic/Latino individuals, which increased from 0.3805 to 0.5265. The precision for Asians or Pacific Islanders in the DT model increased from 0.4727 to 0.6071, and for Non-Hispanic Blacks, it rose from 0.5492 to 0.6657. Gradient Boosting Machines (GBM) produced mixed results, showing accuracy and precision improvements for Non-Hispanic Black and Asian groups, but declines for the Hispanic/Latino and American Indian groups, with the most significant decline in precision, which dropped from 0.4612 to 0.2406 in the American Indian group. Logistic Regression (LR) demonstrated minimal changes across all metrics and groups. For the Non-Hispanic American Indian group, most models showed limited benefits, with several performance metrics either remaining stable or declining.</p><p><strong>Conclusions: </strong>This study demonstrates the potential of AI in predicting COVID-19 mortality while also underscoring the critical need to address demographic biases. The application of transfer learning significantly improved the predictive performance of models across various racial and ethnic groups, suggesting these techniques are effective in mitigating biases and promoting fairness in AI models.</p>","PeriodicalId":9340,"journal":{"name":"BMC Medical Informatics and Decision Making","volume":"25 1","pages":"30"},"PeriodicalIF":3.3000,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11742213/pdf/","citationCount":"0","resultStr":"{\"title\":\"Mitigating bias in AI mortality predictions for minority populations: a transfer learning approach.\",\"authors\":\"Tianshu Gu, Wensen Pan, Jing Yu, Guang Ji, Xia Meng, Yongjun Wang, Minghui Li\",\"doi\":\"10.1186/s12911-025-02862-7\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>The COVID-19 pandemic has highlighted the crucial role of artificial intelligence (AI) in predicting mortality and guiding healthcare decisions. However, AI models may perpetuate or exacerbate existing health disparities due to demographic biases, particularly affecting racial and ethnic minorities. The objective of this study is to investigate the demographic biases in AI models predicting COVID-19 mortality and to assess the effectiveness of transfer learning in improving model fairness across diverse demographic groups.</p><p><strong>Methods: </strong>This retrospective cohort study used a population-based dataset of COVID-19 cases from the Centers for Disease Control and Prevention (CDC), spanning the years 2020-2024. The study analyzed AI model performance across different racial and ethnic groups and employed transfer learning techniques to improve model fairness by adapting pre-trained models to the specific demographic and clinical characteristics of the population.</p><p><strong>Results: </strong>Decision Tree (DT) and Random Forest (RF) models consistently showed improvements in accuracy, precision, and ROC-AUC scores for Non-Hispanic Black, Hispanic/Latino, and Asian populations. The most significant precision improvement was observed in the DT model for Hispanic/Latino individuals, which increased from 0.3805 to 0.5265. The precision for Asians or Pacific Islanders in the DT model increased from 0.4727 to 0.6071, and for Non-Hispanic Blacks, it rose from 0.5492 to 0.6657. Gradient Boosting Machines (GBM) produced mixed results, showing accuracy and precision improvements for Non-Hispanic Black and Asian groups, but declines for the Hispanic/Latino and American Indian groups, with the most significant decline in precision, which dropped from 0.4612 to 0.2406 in the American Indian group. Logistic Regression (LR) demonstrated minimal changes across all metrics and groups. For the Non-Hispanic American Indian group, most models showed limited benefits, with several performance metrics either remaining stable or declining.</p><p><strong>Conclusions: </strong>This study demonstrates the potential of AI in predicting COVID-19 mortality while also underscoring the critical need to address demographic biases. The application of transfer learning significantly improved the predictive performance of models across various racial and ethnic groups, suggesting these techniques are effective in mitigating biases and promoting fairness in AI models.</p>\",\"PeriodicalId\":9340,\"journal\":{\"name\":\"BMC Medical Informatics and Decision Making\",\"volume\":\"25 1\",\"pages\":\"30\"},\"PeriodicalIF\":3.3000,\"publicationDate\":\"2025-01-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11742213/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"BMC Medical Informatics and Decision Making\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1186/s12911-025-02862-7\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"MEDICAL INFORMATICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC Medical Informatics and Decision Making","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1186/s12911-025-02862-7","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MEDICAL INFORMATICS","Score":null,"Total":0}
引用次数: 0

摘要

背景:2019冠状病毒病大流行凸显了人工智能(AI)在预测死亡率和指导医疗保健决策方面的关键作用。然而,由于人口偏见,人工智能模型可能会延续或加剧现有的健康差距,特别是对种族和少数民族的影响。本研究的目的是调查预测COVID-19死亡率的人工智能模型中的人口统计学偏差,并评估迁移学习在提高不同人口统计学群体模型公平性方面的有效性。方法:本回顾性队列研究使用了美国疾病控制与预防中心(CDC)基于人群的COVID-19病例数据集,时间跨度为2020-2024年。该研究分析了人工智能模型在不同种族和民族群体中的表现,并采用迁移学习技术,通过使预训练的模型适应人口的特定人口统计学和临床特征,来提高模型的公平性。结果:决策树(DT)和随机森林(RF)模型一致显示非西班牙裔黑人、西班牙裔/拉丁裔和亚洲人群的准确性、精密度和ROC-AUC得分有所提高。在西班牙裔/拉丁裔个体的DT模型中,精度提高最为显著,从0.3805提高到0.5265。DT模型中亚洲人或太平洋岛民的精度从0.4727提高到0.6071,非西班牙裔黑人的精度从0.5492提高到0.6657。梯度增强机器(GBM)产生了好坏参半的结果,显示非西班牙裔黑人和亚洲群体的准确性和精度提高,但西班牙裔/拉丁裔和美洲印第安人群体的准确性下降,精度下降最显著,从0.4612下降到0.2406在美国印第安人群体。逻辑回归(LR)显示了所有指标和组之间的最小变化。对于非西班牙裔美国印第安人群体,大多数模型显示出有限的好处,几个性能指标要么保持稳定,要么下降。结论:这项研究证明了人工智能在预测COVID-19死亡率方面的潜力,同时也强调了解决人口统计学偏差的迫切需要。迁移学习的应用显著提高了模型在不同种族和民族群体中的预测性能,表明这些技术在减轻偏见和促进人工智能模型的公平性方面是有效的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Mitigating bias in AI mortality predictions for minority populations: a transfer learning approach.

Background: The COVID-19 pandemic has highlighted the crucial role of artificial intelligence (AI) in predicting mortality and guiding healthcare decisions. However, AI models may perpetuate or exacerbate existing health disparities due to demographic biases, particularly affecting racial and ethnic minorities. The objective of this study is to investigate the demographic biases in AI models predicting COVID-19 mortality and to assess the effectiveness of transfer learning in improving model fairness across diverse demographic groups.

Methods: This retrospective cohort study used a population-based dataset of COVID-19 cases from the Centers for Disease Control and Prevention (CDC), spanning the years 2020-2024. The study analyzed AI model performance across different racial and ethnic groups and employed transfer learning techniques to improve model fairness by adapting pre-trained models to the specific demographic and clinical characteristics of the population.

Results: Decision Tree (DT) and Random Forest (RF) models consistently showed improvements in accuracy, precision, and ROC-AUC scores for Non-Hispanic Black, Hispanic/Latino, and Asian populations. The most significant precision improvement was observed in the DT model for Hispanic/Latino individuals, which increased from 0.3805 to 0.5265. The precision for Asians or Pacific Islanders in the DT model increased from 0.4727 to 0.6071, and for Non-Hispanic Blacks, it rose from 0.5492 to 0.6657. Gradient Boosting Machines (GBM) produced mixed results, showing accuracy and precision improvements for Non-Hispanic Black and Asian groups, but declines for the Hispanic/Latino and American Indian groups, with the most significant decline in precision, which dropped from 0.4612 to 0.2406 in the American Indian group. Logistic Regression (LR) demonstrated minimal changes across all metrics and groups. For the Non-Hispanic American Indian group, most models showed limited benefits, with several performance metrics either remaining stable or declining.

Conclusions: This study demonstrates the potential of AI in predicting COVID-19 mortality while also underscoring the critical need to address demographic biases. The application of transfer learning significantly improved the predictive performance of models across various racial and ethnic groups, suggesting these techniques are effective in mitigating biases and promoting fairness in AI models.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
7.20
自引率
5.70%
发文量
297
审稿时长
1 months
期刊介绍: BMC Medical Informatics and Decision Making is an open access journal publishing original peer-reviewed research articles in relation to the design, development, implementation, use, and evaluation of health information technologies and decision-making for human health.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信