Standardization and accuracy of race and ethnicity data: Equity implications for medical AI.

IF 7.7
PLOS digital health Pub Date : 2025-05-29 eCollection Date: 2025-05-01 DOI:10.1371/journal.pdig.0000807
Alexandra Tsalidis, Lakshmi Bharadwaj, Francis X Shen
{"title":"Standardization and accuracy of race and ethnicity data: Equity implications for medical AI.","authors":"Alexandra Tsalidis, Lakshmi Bharadwaj, Francis X Shen","doi":"10.1371/journal.pdig.0000807","DOIUrl":null,"url":null,"abstract":"<p><p>The rapid integration of artificial intelligence (AI) into healthcare has raised many concerns about race bias in AI models. Yet, overlooked in this dialogue is the lack of quality control for the accuracy of patient race and ethnicity (r/e) data in electronic health records (EHR). This article critically examines the factors driving inaccurate and unrepresentative r/e datasets. These include conceptual uncertainties about how to categorize races and ethnicity, shortcomings in data collection practices, EHR standards, and the misclassification of patients' race or ethnicity. To address these challenges, we propose a two-pronged action plan. First, we present a set of best practices for healthcare systems and medical AI researchers to improve r/e data accuracy. Second, we call for developers of medical AI models to transparently warrant the quality of their r/e data. Given the ethical and scientific imperatives of ensuring high-quality r/e data in AI-driven healthcare, we argue that these steps should be taken immediately.</p>","PeriodicalId":74465,"journal":{"name":"PLOS digital health","volume":"4 5","pages":"e0000807"},"PeriodicalIF":7.7000,"publicationDate":"2025-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12121746/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"PLOS digital health","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1371/journal.pdig.0000807","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/5/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The rapid integration of artificial intelligence (AI) into healthcare has raised many concerns about race bias in AI models. Yet, overlooked in this dialogue is the lack of quality control for the accuracy of patient race and ethnicity (r/e) data in electronic health records (EHR). This article critically examines the factors driving inaccurate and unrepresentative r/e datasets. These include conceptual uncertainties about how to categorize races and ethnicity, shortcomings in data collection practices, EHR standards, and the misclassification of patients' race or ethnicity. To address these challenges, we propose a two-pronged action plan. First, we present a set of best practices for healthcare systems and medical AI researchers to improve r/e data accuracy. Second, we call for developers of medical AI models to transparently warrant the quality of their r/e data. Given the ethical and scientific imperatives of ensuring high-quality r/e data in AI-driven healthcare, we argue that these steps should be taken immediately.

种族和民族数据的标准化和准确性:对医疗人工智能的公平影响。
人工智能(AI)与医疗保健的快速融合引发了许多人对人工智能模型中种族偏见的担忧。然而,在这种对话中被忽视的是,缺乏对电子健康记录(EHR)中患者种族和族裔(r/e)数据准确性的质量控制。本文批判性地考察了导致不准确和不具代表性的r/e数据集的因素。这些包括关于如何对种族和民族进行分类的概念上的不确定性,数据收集实践中的缺陷,电子病历标准,以及对患者种族或民族的错误分类。为应对这些挑战,我们提出一项双管齐下的行动计划。首先,我们为医疗保健系统和医疗人工智能研究人员提供了一套最佳实践,以提高r/e数据的准确性。其次,我们呼吁医疗人工智能模型的开发人员透明地保证其r/e数据的质量。鉴于在人工智能驱动的医疗保健中确保高质量r/e数据的伦理和科学必要性,我们认为应立即采取这些步骤。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信