Pitfalls in Interpretive Applications of Artificial Intelligence in Radiology.

IF 4.7 2区 医学 Q1 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING
Shima Behzad, Seyed Mohammad Hossein Tabatabaei, Max Yang Lu, Liesl S Eibschutz, Ali Gholamrezanezhad
{"title":"Pitfalls in Interpretive Applications of Artificial Intelligence in Radiology.","authors":"Shima Behzad, Seyed Mohammad Hossein Tabatabaei, Max Yang Lu, Liesl S Eibschutz, Ali Gholamrezanezhad","doi":"10.2214/AJR.24.31493","DOIUrl":null,"url":null,"abstract":"<p><p>Interpretive artificial intelligence (AI) tools are poised to change the future of radiology. However, certain pitfalls may pose particular challenges for optimal AI interpretative performance. These include anatomic variants, age-related changes, postoperative changes, medical devices, image artifacts, lack of integration of prior and concurrent imaging examinations and clinical information, as well as the satisfaction-of-search effect. Model training and development should account for such pitfalls, to minimize errors and optimize interpretation accuracy. More broadly, AI algorithms should be exposed to diverse and complex training data sets, to yield a holistic interpretation that considers all relevant information beyond the individual examination. Successful clinical deployment of AI tools will require that radiologist end-users recognize these pitfalls and other limitations of the available models. Furthermore, developers should incorporate explainable AI techniques (e.g., heat maps) into their tools, to improve radiologists' understanding of model outputs and to enable radiologists to provide feedback for guiding continuous learning and iterative refinement. In this article, we provide an overview of common pitfalls that radiologists may encounter when using interpretive AI products in daily practice. We present how such pitfalls lead to AI errors and offer potential strategies that AI developers may use for their mitigation.</p>","PeriodicalId":55529,"journal":{"name":"American Journal of Roentgenology","volume":null,"pages":null},"PeriodicalIF":4.7000,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"American Journal of Roentgenology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.2214/AJR.24.31493","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

Abstract

Interpretive artificial intelligence (AI) tools are poised to change the future of radiology. However, certain pitfalls may pose particular challenges for optimal AI interpretative performance. These include anatomic variants, age-related changes, postoperative changes, medical devices, image artifacts, lack of integration of prior and concurrent imaging examinations and clinical information, as well as the satisfaction-of-search effect. Model training and development should account for such pitfalls, to minimize errors and optimize interpretation accuracy. More broadly, AI algorithms should be exposed to diverse and complex training data sets, to yield a holistic interpretation that considers all relevant information beyond the individual examination. Successful clinical deployment of AI tools will require that radiologist end-users recognize these pitfalls and other limitations of the available models. Furthermore, developers should incorporate explainable AI techniques (e.g., heat maps) into their tools, to improve radiologists' understanding of model outputs and to enable radiologists to provide feedback for guiding continuous learning and iterative refinement. In this article, we provide an overview of common pitfalls that radiologists may encounter when using interpretive AI products in daily practice. We present how such pitfalls lead to AI errors and offer potential strategies that AI developers may use for their mitigation.

放射学中人工智能应用的误区。
人工智能(AI)判读工具有望改变放射学的未来。然而,某些隐患可能会对人工智能的最佳解释性能构成特殊挑战。其中包括解剖变异、年龄相关变化、术后变化、医疗设备、图像伪影、缺乏对先前和同期成像检查和临床信息的整合,以及搜索满意度效应。模型的训练和开发应考虑到这些隐患,以尽量减少误差并优化解释的准确性。更广泛地说,人工智能算法应接触多样化和复杂的训练数据集,以产生考虑到单项检查以外所有相关信息的整体解释。要想在临床上成功应用人工智能工具,放射科医生的最终用户就必须认识到这些缺陷以及现有模型的其他局限性。此外,开发人员应将可解释的人工智能技术(如热图)纳入其工具,以提高放射科医生对模型输出的理解,并使放射科医生能够提供反馈,以指导持续学习和迭代改进。在本文中,我们将概述放射科医生在日常实践中使用解释性人工智能产品时可能会遇到的常见陷阱。我们介绍了这些陷阱是如何导致人工智能错误的,并提供了人工智能开发人员可用于缓解这些错误的潜在策略。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
12.80
自引率
4.00%
发文量
920
审稿时长
3 months
期刊介绍: Founded in 1907, the monthly American Journal of Roentgenology (AJR) is the world’s longest continuously published general radiology journal. AJR is recognized as among the specialty’s leading peer-reviewed journals and has a worldwide circulation of close to 25,000. The journal publishes clinically-oriented articles across all radiology subspecialties, seeking relevance to radiologists’ daily practice. The journal publishes hundreds of articles annually with a diverse range of formats, including original research, reviews, clinical perspectives, editorials, and other short reports. The journal engages its audience through a spectrum of social media and digital communication activities.
文献相关原料
公司名称 产品信息 采购帮参考价格
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信