连接临床知识和人工智能在胸部放射学中的可解释性

iRadiology Pub Date : 2025-06-25 DOI:10.1002/ird3.70015
Mengze Xu
{"title":"连接临床知识和人工智能在胸部放射学中的可解释性","authors":"Mengze Xu","doi":"10.1002/ird3.70015","DOIUrl":null,"url":null,"abstract":"<p>Yuan's study [<span>1</span>] entitled “<i>Anatomic Boundary-Aware Explanation for Convolutional Neural Networks in Diagnostic Radiology</i>” underscores a fundamental gap in existing XAI approaches: the neglect of clinical domain knowledge. Thoracic diseases primarily manifest within specific anatomical regions, such as the lung parenchyma. Yet, conventional XAI methods such as Grad-CAM or Integrated Gradients often highlight extraneous areas (e.g., medical devices, chest wall artifacts), leading to misinterpretations. By leveraging anatomic boundaries derived from a pretrained lung segmentation model, the authors enforce spatial constraints on CNN explanations, aligning them with clinically relevant regions. This innovation is particularly impactful for resource-limited settings, where annotations for fine-grained lesion localization are scarce.</p><p>The study's quantitative results are compelling: Across 72 scenarios involving 3 CNN architectures, 4 diseases, and 2 classification settings, the boundary-aware method outperformed baseline explanations in 71 cases. For example, in pneumothorax detection, the dice similarity coefficient (DSC) improved by up to 5.09% when integrating anatomic constraints. These findings validate the hypothesis that incorporating radiological expertise into XAI frameworks enhances explanation fidelity.</p><p>The paper's strengths lie in its plug-and-play design and transfer learning strategy. By decoupling lung segmentation from the CNN classifier, the authors avoid retraining on annotated target datasets, reducing computational and labeling costs. The use of publicly available segmentation datasets (e.g., Japanese Society of Radiological Technology) ensures reproducibility and scalability. However, this approach assumes minimal domain shift between external and target datasets. Future studies should evaluate robustness across diverse imaging protocols or patient populations, where anatomical variations (e.g., emphysematous lungs, postsurgical changes) might affect segmentation accuracy. Another notable aspect is the comprehensive evaluation of multiple XAI methods (saliency map, Grad-CAM, Integrated Gradients) and CNN architectures (VGG-11, ResNet-18, AlexNet) [<span>2</span>]. The consistent improvements observed across these configurations suggest the boundary-aware framework is generalizable. However, the reliance on lightweight CNNs (e.g., VGG-11) raises questions about applicability to modern, deeper models (e.g., vision transformers), which may require different regularization strategies.</p><p>A limitation is the qualitative gap between improved metrics and clinical utility. Although intersection over union and DSC metrics quantify overlap with ground-truth lesions, they do not directly measure radiologists' trust in AI explanations. Future work should incorporate human-in-the-loop studies to assess how boundary-aware explanations influence diagnostic decisions and workflow efficiency.</p><p>Yuan's approach opens new avenues for integrating domain knowledge into XAI. For instance, extending anatomical constraints to other organs (e.g., heart, mediastinum) could enhance explanations for complex pathologies such as aortic aneurysms. Additionally, combining boundary-aware XAI with weakly supervised learning might improve lesion segmentation accuracy, addressing the study's low DSC values (e.g., &lt; 10% for certain mass explanations).</p><p>The paper also highlights the tension between multilabel and binary classification. Although binary classifiers showed superior explanation performance, clinical practice often demands multilabel predictions. Future research could explore hybrid approaches, such as using binary classifiers as building blocks for multilabel tasks, as proposed by Shiraishi and Fukumizu [<span>3</span>].</p><p>Advanced segmentation models: Incorporate state-of-the-art segmentation tools such as MedSAM [<span>4</span>] to improve boundary precision, especially in challenging cases. Dynamic constraints: Develop disease-specific boundaries (e.g., differentiating pneumothorax from atelectasis) to refine explanations further. Real-world validation: Conduct randomized controlled trials to evaluate how boundary-aware explanations affect radiologists' diagnostic accuracy and confidence. Generalization to other modalities: Adapt the framework to computed tomography or magnetic resonance imaging, where organ segmentation is equally critical.</p><p>Yuan's study represents a pivotal step toward bridging clinical knowledge and AI interpretability in thoracic radiology. By constraining CNN explanations to anatomical boundaries, the authors demonstrate that domain-specific regularization can mitigate shortcut learning and align AI reasoning with clinical intuition. Although challenges remain—including validation in diverse populations and integration with advanced models—the proposed framework sets a precedent for knowledge-driven XAI in medical imaging. As AI transitions from research to clinical practice, such innovations will be essential for fostering trust and ensuring safe, effective patient care.</p><p><b>Mengze Xu:</b> conceptualization (lead), investigation (lead), supervision (lead), writing – original draft (lead).</p><p>The author has nothing to report.</p><p>The author has nothing to report.</p><p>The author declares no conflicts of interest.</p>","PeriodicalId":73508,"journal":{"name":"iRadiology","volume":"3 4","pages":"311-312"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ird3.70015","citationCount":"0","resultStr":"{\"title\":\"Bridging Clinical Knowledge and AI Interpretability in Thoracic Radiology\",\"authors\":\"Mengze Xu\",\"doi\":\"10.1002/ird3.70015\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Yuan's study [<span>1</span>] entitled “<i>Anatomic Boundary-Aware Explanation for Convolutional Neural Networks in Diagnostic Radiology</i>” underscores a fundamental gap in existing XAI approaches: the neglect of clinical domain knowledge. Thoracic diseases primarily manifest within specific anatomical regions, such as the lung parenchyma. Yet, conventional XAI methods such as Grad-CAM or Integrated Gradients often highlight extraneous areas (e.g., medical devices, chest wall artifacts), leading to misinterpretations. By leveraging anatomic boundaries derived from a pretrained lung segmentation model, the authors enforce spatial constraints on CNN explanations, aligning them with clinically relevant regions. This innovation is particularly impactful for resource-limited settings, where annotations for fine-grained lesion localization are scarce.</p><p>The study's quantitative results are compelling: Across 72 scenarios involving 3 CNN architectures, 4 diseases, and 2 classification settings, the boundary-aware method outperformed baseline explanations in 71 cases. For example, in pneumothorax detection, the dice similarity coefficient (DSC) improved by up to 5.09% when integrating anatomic constraints. These findings validate the hypothesis that incorporating radiological expertise into XAI frameworks enhances explanation fidelity.</p><p>The paper's strengths lie in its plug-and-play design and transfer learning strategy. By decoupling lung segmentation from the CNN classifier, the authors avoid retraining on annotated target datasets, reducing computational and labeling costs. The use of publicly available segmentation datasets (e.g., Japanese Society of Radiological Technology) ensures reproducibility and scalability. However, this approach assumes minimal domain shift between external and target datasets. Future studies should evaluate robustness across diverse imaging protocols or patient populations, where anatomical variations (e.g., emphysematous lungs, postsurgical changes) might affect segmentation accuracy. Another notable aspect is the comprehensive evaluation of multiple XAI methods (saliency map, Grad-CAM, Integrated Gradients) and CNN architectures (VGG-11, ResNet-18, AlexNet) [<span>2</span>]. The consistent improvements observed across these configurations suggest the boundary-aware framework is generalizable. However, the reliance on lightweight CNNs (e.g., VGG-11) raises questions about applicability to modern, deeper models (e.g., vision transformers), which may require different regularization strategies.</p><p>A limitation is the qualitative gap between improved metrics and clinical utility. Although intersection over union and DSC metrics quantify overlap with ground-truth lesions, they do not directly measure radiologists' trust in AI explanations. Future work should incorporate human-in-the-loop studies to assess how boundary-aware explanations influence diagnostic decisions and workflow efficiency.</p><p>Yuan's approach opens new avenues for integrating domain knowledge into XAI. For instance, extending anatomical constraints to other organs (e.g., heart, mediastinum) could enhance explanations for complex pathologies such as aortic aneurysms. Additionally, combining boundary-aware XAI with weakly supervised learning might improve lesion segmentation accuracy, addressing the study's low DSC values (e.g., &lt; 10% for certain mass explanations).</p><p>The paper also highlights the tension between multilabel and binary classification. Although binary classifiers showed superior explanation performance, clinical practice often demands multilabel predictions. Future research could explore hybrid approaches, such as using binary classifiers as building blocks for multilabel tasks, as proposed by Shiraishi and Fukumizu [<span>3</span>].</p><p>Advanced segmentation models: Incorporate state-of-the-art segmentation tools such as MedSAM [<span>4</span>] to improve boundary precision, especially in challenging cases. Dynamic constraints: Develop disease-specific boundaries (e.g., differentiating pneumothorax from atelectasis) to refine explanations further. Real-world validation: Conduct randomized controlled trials to evaluate how boundary-aware explanations affect radiologists' diagnostic accuracy and confidence. Generalization to other modalities: Adapt the framework to computed tomography or magnetic resonance imaging, where organ segmentation is equally critical.</p><p>Yuan's study represents a pivotal step toward bridging clinical knowledge and AI interpretability in thoracic radiology. By constraining CNN explanations to anatomical boundaries, the authors demonstrate that domain-specific regularization can mitigate shortcut learning and align AI reasoning with clinical intuition. Although challenges remain—including validation in diverse populations and integration with advanced models—the proposed framework sets a precedent for knowledge-driven XAI in medical imaging. As AI transitions from research to clinical practice, such innovations will be essential for fostering trust and ensuring safe, effective patient care.</p><p><b>Mengze Xu:</b> conceptualization (lead), investigation (lead), supervision (lead), writing – original draft (lead).</p><p>The author has nothing to report.</p><p>The author has nothing to report.</p><p>The author declares no conflicts of interest.</p>\",\"PeriodicalId\":73508,\"journal\":{\"name\":\"iRadiology\",\"volume\":\"3 4\",\"pages\":\"311-312\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-06-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ird3.70015\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"iRadiology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/ird3.70015\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"iRadiology","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/ird3.70015","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

Yuan的研究标题为“诊断放射学中卷积神经网络的解剖边界感知解释”,强调了现有XAI方法的根本缺陷:忽视临床领域知识。胸部疾病主要表现在特定的解剖区域,如肺实质。然而,传统的XAI方法,如Grad-CAM或Integrated Gradients,通常会突出无关区域(例如,医疗设备、胸壁伪影),导致误解。通过利用来自预训练的肺分割模型的解剖边界,作者对CNN解释施加空间约束,使其与临床相关区域保持一致。这种创新对资源有限的环境特别有影响,在这种环境中,对细粒度病变定位的注释很少。该研究的定量结果令人信服:在涉及3种CNN架构、4种疾病和2种分类设置的72种场景中,边界感知方法在71种情况下优于基线解释。例如,在气胸检测中,当整合解剖约束时,骰子相似系数(DSC)提高了5.09%。这些发现验证了将放射学专业知识纳入XAI框架可以提高解释保真度的假设。本文的优势在于即插即用设计和迁移学习策略。通过将肺分割与CNN分类器解耦,作者避免了对带注释的目标数据集进行再训练,减少了计算和标记成本。使用公开可用的分割数据集(例如,日本放射技术学会)确保可重复性和可扩展性。然而,这种方法假设外部和目标数据集之间的域转移最小。未来的研究应评估不同成像方案或患者群体的稳健性,其中解剖差异(例如肺气肿肺,术后改变)可能影响分割准确性。另一个值得注意的方面是对多种XAI方法(saliency map, Grad-CAM, Integrated Gradients)和CNN架构(VGG-11, ResNet-18, AlexNet)[2]的综合评估。在这些配置中观察到的一致改进表明,边界感知框架是可推广的。然而,对轻量级cnn(例如,VGG-11)的依赖引发了对现代深层模型(例如,视觉变压器)适用性的问题,这可能需要不同的正则化策略。一个限制是改进的指标和临床效用之间的定性差距。虽然联合交叉点和DSC指标量化了与真实病灶的重叠,但它们并不能直接衡量放射科医生对人工智能解释的信任。未来的工作应纳入人在循环研究,以评估边界感知解释如何影响诊断决策和工作流程效率。Yuan的方法为将领域知识集成到XAI中开辟了新的途径。例如,将解剖学限制扩展到其他器官(如心脏、纵隔)可以加强对复杂病理(如主动脉瘤)的解释。此外,将边界感知的XAI与弱监督学习相结合可能会提高病灶分割的准确性,解决该研究的低DSC值(例如,对于某些质量解释,&lt; 10%)。本文还强调了多标签和二元分类之间的紧张关系。虽然二元分类器表现出优越的解释性能,但临床实践往往需要多标签预测。未来的研究可以探索混合方法,例如使用二元分类器作为多标签任务的构建块,正如白石和Fukumizu[3]所提出的那样。先进的分割模型:结合最先进的分割工具,如MedSAM[4],以提高边界精度,特别是在具有挑战性的情况下。动态约束:发展疾病特异性界限(例如,区分气胸和肺不张)以进一步完善解释。真实世界验证:进行随机对照试验,以评估边界感知解释如何影响放射科医生的诊断准确性和信心。推广到其他模式:使框架适用于计算机断层扫描或磁共振成像,其中器官分割同样重要。Yuan的研究代表了胸放射学临床知识和人工智能可解释性之间桥梁的关键一步。通过将CNN的解释限制在解剖边界上,作者证明了特定领域的正则化可以缓解快速学习,并使人工智能推理与临床直觉保持一致。尽管挑战依然存在——包括在不同人群中的验证和与先进模型的整合——但所提出的框架为医学成像中知识驱动的XAI树立了先例。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Bridging Clinical Knowledge and AI Interpretability in Thoracic Radiology

Yuan's study [1] entitled “Anatomic Boundary-Aware Explanation for Convolutional Neural Networks in Diagnostic Radiology” underscores a fundamental gap in existing XAI approaches: the neglect of clinical domain knowledge. Thoracic diseases primarily manifest within specific anatomical regions, such as the lung parenchyma. Yet, conventional XAI methods such as Grad-CAM or Integrated Gradients often highlight extraneous areas (e.g., medical devices, chest wall artifacts), leading to misinterpretations. By leveraging anatomic boundaries derived from a pretrained lung segmentation model, the authors enforce spatial constraints on CNN explanations, aligning them with clinically relevant regions. This innovation is particularly impactful for resource-limited settings, where annotations for fine-grained lesion localization are scarce.

The study's quantitative results are compelling: Across 72 scenarios involving 3 CNN architectures, 4 diseases, and 2 classification settings, the boundary-aware method outperformed baseline explanations in 71 cases. For example, in pneumothorax detection, the dice similarity coefficient (DSC) improved by up to 5.09% when integrating anatomic constraints. These findings validate the hypothesis that incorporating radiological expertise into XAI frameworks enhances explanation fidelity.

The paper's strengths lie in its plug-and-play design and transfer learning strategy. By decoupling lung segmentation from the CNN classifier, the authors avoid retraining on annotated target datasets, reducing computational and labeling costs. The use of publicly available segmentation datasets (e.g., Japanese Society of Radiological Technology) ensures reproducibility and scalability. However, this approach assumes minimal domain shift between external and target datasets. Future studies should evaluate robustness across diverse imaging protocols or patient populations, where anatomical variations (e.g., emphysematous lungs, postsurgical changes) might affect segmentation accuracy. Another notable aspect is the comprehensive evaluation of multiple XAI methods (saliency map, Grad-CAM, Integrated Gradients) and CNN architectures (VGG-11, ResNet-18, AlexNet) [2]. The consistent improvements observed across these configurations suggest the boundary-aware framework is generalizable. However, the reliance on lightweight CNNs (e.g., VGG-11) raises questions about applicability to modern, deeper models (e.g., vision transformers), which may require different regularization strategies.

A limitation is the qualitative gap between improved metrics and clinical utility. Although intersection over union and DSC metrics quantify overlap with ground-truth lesions, they do not directly measure radiologists' trust in AI explanations. Future work should incorporate human-in-the-loop studies to assess how boundary-aware explanations influence diagnostic decisions and workflow efficiency.

Yuan's approach opens new avenues for integrating domain knowledge into XAI. For instance, extending anatomical constraints to other organs (e.g., heart, mediastinum) could enhance explanations for complex pathologies such as aortic aneurysms. Additionally, combining boundary-aware XAI with weakly supervised learning might improve lesion segmentation accuracy, addressing the study's low DSC values (e.g., < 10% for certain mass explanations).

The paper also highlights the tension between multilabel and binary classification. Although binary classifiers showed superior explanation performance, clinical practice often demands multilabel predictions. Future research could explore hybrid approaches, such as using binary classifiers as building blocks for multilabel tasks, as proposed by Shiraishi and Fukumizu [3].

Advanced segmentation models: Incorporate state-of-the-art segmentation tools such as MedSAM [4] to improve boundary precision, especially in challenging cases. Dynamic constraints: Develop disease-specific boundaries (e.g., differentiating pneumothorax from atelectasis) to refine explanations further. Real-world validation: Conduct randomized controlled trials to evaluate how boundary-aware explanations affect radiologists' diagnostic accuracy and confidence. Generalization to other modalities: Adapt the framework to computed tomography or magnetic resonance imaging, where organ segmentation is equally critical.

Yuan's study represents a pivotal step toward bridging clinical knowledge and AI interpretability in thoracic radiology. By constraining CNN explanations to anatomical boundaries, the authors demonstrate that domain-specific regularization can mitigate shortcut learning and align AI reasoning with clinical intuition. Although challenges remain—including validation in diverse populations and integration with advanced models—the proposed framework sets a precedent for knowledge-driven XAI in medical imaging. As AI transitions from research to clinical practice, such innovations will be essential for fostering trust and ensuring safe, effective patient care.

Mengze Xu: conceptualization (lead), investigation (lead), supervision (lead), writing – original draft (lead).

The author has nothing to report.

The author has nothing to report.

The author declares no conflicts of interest.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信