Which explanations do clinicians prefer? A comparative evaluation of XAI understandability and actionability in predicting the need for hospitalization.
Laura Bergomi, Giovanna Nicora, Marta Anna Orlowska, Chiara Podrecca, Riccardo Bellazzi, Caterina Fregosi, Francesco Salinaro, Marco Bonzano, Giuseppe Crescenzi, Francesco Speciale, Santi Di Pietro, Valentina Zuccaro, Erika Asperges, Paolo Sacchi, Pietro Valsecchi, Elisabetta Pagani, Michele Catalano, Chandra Bortolotto, Lorenzo Preda, Enea Parimbelli
{"title":"Which explanations do clinicians prefer? A comparative evaluation of XAI understandability and actionability in predicting the need for hospitalization.","authors":"Laura Bergomi, Giovanna Nicora, Marta Anna Orlowska, Chiara Podrecca, Riccardo Bellazzi, Caterina Fregosi, Francesco Salinaro, Marco Bonzano, Giuseppe Crescenzi, Francesco Speciale, Santi Di Pietro, Valentina Zuccaro, Erika Asperges, Paolo Sacchi, Pietro Valsecchi, Elisabetta Pagani, Michele Catalano, Chandra Bortolotto, Lorenzo Preda, Enea Parimbelli","doi":"10.1186/s12911-025-03045-0","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>This study aims to address the gap in understanding clinicians' attitudes toward explainable AI (XAI) methods applied to machine learning models using tabular data, commonly found in clinical settings. It specifically explores clinicians' perceptions of different XAI methods from the ALFABETO project, which predicts COVID-19 patient hospitalization based on clinical, laboratory, and chest X-ray at time of presentation to the Emergency Department. The focus is on two cognitive dimensions: understandability and actionability of the explanations provided by explainable-by-design and post-hoc methods.</p><p><strong>Methods: </strong>A questionnaire-based experiment was conducted with 10 clinicians from the IRCCS Policlinico San Matteo Foundation in Pavia, Italy. Each clinician evaluated 10 real-world cases, rating predictions and explanations from three XAI tools: Bayesian networks, SHapley Additive exPlanations (SHAP), and AraucanaXAI. Two cognitive statements for each method were rated on a Likert scale, as well as the agreement with the prediction. Two clinicians answered the survey during think-aloud interviews.</p><p><strong>Results: </strong>Clinicians demonstrated generally positive attitudes toward AI, but high compliance rates (86% on average) indicate a risk of automation bias. Understandability and actionability are positively correlated, with SHAP being the preferred method due to its simplicity. However, the perception of methods varies according to specialty and expertise.</p><p><strong>Conclusions: </strong>The findings suggest that SHAP and AraucanaXAI are promising candidates for improving the use of XAI in clinical decision support systems (DSSs), highlighting the importance of clinicians' expertise, specialty, and setting on the selection and development of supportive XAI advice. Finally, the study provides valuable insights into the design of future XAI DSSs.</p>","PeriodicalId":9340,"journal":{"name":"BMC Medical Informatics and Decision Making","volume":"25 1","pages":"269"},"PeriodicalIF":3.3000,"publicationDate":"2025-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12265231/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC Medical Informatics and Decision Making","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1186/s12911-025-03045-0","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MEDICAL INFORMATICS","Score":null,"Total":0}
引用次数: 0
Abstract
Background: This study aims to address the gap in understanding clinicians' attitudes toward explainable AI (XAI) methods applied to machine learning models using tabular data, commonly found in clinical settings. It specifically explores clinicians' perceptions of different XAI methods from the ALFABETO project, which predicts COVID-19 patient hospitalization based on clinical, laboratory, and chest X-ray at time of presentation to the Emergency Department. The focus is on two cognitive dimensions: understandability and actionability of the explanations provided by explainable-by-design and post-hoc methods.
Methods: A questionnaire-based experiment was conducted with 10 clinicians from the IRCCS Policlinico San Matteo Foundation in Pavia, Italy. Each clinician evaluated 10 real-world cases, rating predictions and explanations from three XAI tools: Bayesian networks, SHapley Additive exPlanations (SHAP), and AraucanaXAI. Two cognitive statements for each method were rated on a Likert scale, as well as the agreement with the prediction. Two clinicians answered the survey during think-aloud interviews.
Results: Clinicians demonstrated generally positive attitudes toward AI, but high compliance rates (86% on average) indicate a risk of automation bias. Understandability and actionability are positively correlated, with SHAP being the preferred method due to its simplicity. However, the perception of methods varies according to specialty and expertise.
Conclusions: The findings suggest that SHAP and AraucanaXAI are promising candidates for improving the use of XAI in clinical decision support systems (DSSs), highlighting the importance of clinicians' expertise, specialty, and setting on the selection and development of supportive XAI advice. Finally, the study provides valuable insights into the design of future XAI DSSs.
背景:本研究旨在解决临床医生对可解释人工智能(XAI)方法的态度差异,这些方法应用于使用表格数据的机器学习模型,通常在临床环境中发现。它专门探讨了临床医生对ALFABETO项目中不同XAI方法的看法,该项目根据临床、实验室和急诊时的胸部x光片预测COVID-19患者的住院情况。重点是两个认知维度:可解释的设计和事后方法提供的解释的可理解性和可操作性。方法:对意大利帕维亚市IRCCS polilinico San Matteo基金会的10名临床医生进行问卷调查。每位临床医生评估了10个真实世界的病例,对三种XAI工具的预测和解释进行了评级:贝叶斯网络、SHapley加性解释(SHAP)和AraucanaXAI。每种方法的两个认知陈述都用李克特量表进行评分,以及与预测的一致性。两位临床医生在畅想访谈中回答了这项调查。结果:临床医生普遍对人工智能持积极态度,但高依从率(平均86%)表明存在自动化偏见的风险。可理解性和可操作性正相关,SHAP因其简单性而成为首选方法。然而,对方法的看法因专业和专业知识而异。结论:研究结果表明,SHAP和AraucanaXAI是改善临床决策支持系统(DSSs)中XAI使用的有希望的候选者,突出了临床医生的专业知识、专业和环境对支持性XAI建议的选择和发展的重要性。最后,该研究为未来XAI dss的设计提供了有价值的见解。
期刊介绍:
BMC Medical Informatics and Decision Making is an open access journal publishing original peer-reviewed research articles in relation to the design, development, implementation, use, and evaluation of health information technologies and decision-making for human health.