补充可解释的机器学习与协同分析策略,用于甲状腺癌复发预测

IF 3.2 3区 医学 Q1 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING
Souichi Oka , Yoshiyasu Takefuji
{"title":"补充可解释的机器学习与协同分析策略,用于甲状腺癌复发预测","authors":"Souichi Oka ,&nbsp;Yoshiyasu Takefuji","doi":"10.1016/j.ejrad.2025.112308","DOIUrl":null,"url":null,"abstract":"<div><div>This correspondence critically examines the methodology of Schindele et al. (2025) on thyroid cancer recurrence prediction. While their interpretable XGBoost model achieved a high predictive accuracy of 95.8% and a 0.947 AUROC, it is crucial to recognize that this predictive power does not justify the reliability of its derived feature importance rankings. As widely acknowledged in the literature, high predictive accuracy does not guarantee unbiased or reliable feature attribution. We underscore that gradient boosting decision tree (GBDT) models, including XGBoost, are prone to inherent biases in feature importance estimation, often due to overfitting. Furthermore, SHapley Additive exPlanations (SHAP), a widely adopted explainable AI (XAI) technique, can inherit and even amplify these biases, given its model-dependent nature. This raises concerns about the interpretive validity of the identified risk factors. To mitigate these methodological limitations, we advocate for integrative analytical frameworks that combine machine learning with robust statistical and non-parametric approaches, such as Highly Variable Feature Selection (HVFS) and Independent Component Analysis (ICA). These multi-faceted strategies are indispensable for obtaining robust and interpretable insights into feature importance, warranting their prioritization in future research efforts.</div></div>","PeriodicalId":12063,"journal":{"name":"European Journal of Radiology","volume":"191 ","pages":"Article 112308"},"PeriodicalIF":3.2000,"publicationDate":"2025-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Complementing interpretable machine learning with synergistic analytical strategies for thyroid cancer recurrence prediction\",\"authors\":\"Souichi Oka ,&nbsp;Yoshiyasu Takefuji\",\"doi\":\"10.1016/j.ejrad.2025.112308\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This correspondence critically examines the methodology of Schindele et al. (2025) on thyroid cancer recurrence prediction. While their interpretable XGBoost model achieved a high predictive accuracy of 95.8% and a 0.947 AUROC, it is crucial to recognize that this predictive power does not justify the reliability of its derived feature importance rankings. As widely acknowledged in the literature, high predictive accuracy does not guarantee unbiased or reliable feature attribution. We underscore that gradient boosting decision tree (GBDT) models, including XGBoost, are prone to inherent biases in feature importance estimation, often due to overfitting. Furthermore, SHapley Additive exPlanations (SHAP), a widely adopted explainable AI (XAI) technique, can inherit and even amplify these biases, given its model-dependent nature. This raises concerns about the interpretive validity of the identified risk factors. To mitigate these methodological limitations, we advocate for integrative analytical frameworks that combine machine learning with robust statistical and non-parametric approaches, such as Highly Variable Feature Selection (HVFS) and Independent Component Analysis (ICA). These multi-faceted strategies are indispensable for obtaining robust and interpretable insights into feature importance, warranting their prioritization in future research efforts.</div></div>\",\"PeriodicalId\":12063,\"journal\":{\"name\":\"European Journal of Radiology\",\"volume\":\"191 \",\"pages\":\"Article 112308\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-07-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"European Journal of Radiology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0720048X25003948\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"European Journal of Radiology","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0720048X25003948","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

摘要

本文对Schindele等人(2025)关于甲状腺癌复发预测的方法进行了批判性的研究。虽然他们的可解释的XGBoost模型实现了95.8%的高预测精度和0.947 AUROC,但重要的是要认识到,这种预测能力并不能证明其导出的特征重要性排名的可靠性。正如文献中广泛承认的那样,高预测准确度并不能保证无偏或可靠的特征归因。我们强调,梯度增强决策树(GBDT)模型,包括XGBoost,在特征重要性估计中容易产生固有偏差,通常是由于过拟合。此外,SHapley加性解释(SHAP),一种被广泛采用的可解释人工智能(XAI)技术,可以继承甚至放大这些偏差,因为它具有模型依赖的性质。这引起了对已确定的风险因素的解释有效性的关注。为了减轻这些方法上的限制,我们提倡将机器学习与鲁棒统计和非参数方法相结合的综合分析框架,如高变量特征选择(HVFS)和独立成分分析(ICA)。这些多方面的策略对于获得对特征重要性的强大和可解释的见解是必不可少的,保证了它们在未来研究工作中的优先级。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Complementing interpretable machine learning with synergistic analytical strategies for thyroid cancer recurrence prediction
This correspondence critically examines the methodology of Schindele et al. (2025) on thyroid cancer recurrence prediction. While their interpretable XGBoost model achieved a high predictive accuracy of 95.8% and a 0.947 AUROC, it is crucial to recognize that this predictive power does not justify the reliability of its derived feature importance rankings. As widely acknowledged in the literature, high predictive accuracy does not guarantee unbiased or reliable feature attribution. We underscore that gradient boosting decision tree (GBDT) models, including XGBoost, are prone to inherent biases in feature importance estimation, often due to overfitting. Furthermore, SHapley Additive exPlanations (SHAP), a widely adopted explainable AI (XAI) technique, can inherit and even amplify these biases, given its model-dependent nature. This raises concerns about the interpretive validity of the identified risk factors. To mitigate these methodological limitations, we advocate for integrative analytical frameworks that combine machine learning with robust statistical and non-parametric approaches, such as Highly Variable Feature Selection (HVFS) and Independent Component Analysis (ICA). These multi-faceted strategies are indispensable for obtaining robust and interpretable insights into feature importance, warranting their prioritization in future research efforts.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
6.70
自引率
3.00%
发文量
398
审稿时长
42 days
期刊介绍: European Journal of Radiology is an international journal which aims to communicate to its readers, state-of-the-art information on imaging developments in the form of high quality original research articles and timely reviews on current developments in the field. Its audience includes clinicians at all levels of training including radiology trainees, newly qualified imaging specialists and the experienced radiologist. Its aim is to inform efficient, appropriate and evidence-based imaging practice to the benefit of patients worldwide.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信