Improving disease classification performance and explainability of deep learning models in radiology with heatmap generators.

Akino Watanabe, Sara Ketabi, Khashayar Namdar, Farzad Khalvati
{"title":"Improving disease classification performance and explainability of deep learning models in radiology with heatmap generators.","authors":"Akino Watanabe,&nbsp;Sara Ketabi,&nbsp;Khashayar Namdar,&nbsp;Farzad Khalvati","doi":"10.3389/fradi.2022.991683","DOIUrl":null,"url":null,"abstract":"<p><p>As deep learning is widely used in the radiology field, the explainability of Artificial Intelligence (AI) models is becoming increasingly essential to gain clinicians' trust when using the models for diagnosis. In this research, three experiment sets were conducted with a U-Net architecture to improve the disease classification performance while enhancing the heatmaps corresponding to the model's focus through incorporating heatmap generators during training. All experiments used the dataset that contained chest radiographs, associated labels from one of the three conditions [\"normal\", \"congestive heart failure (CHF)\", and \"pneumonia\"], and numerical information regarding a radiologist's eye-gaze coordinates on the images. The paper that introduced this dataset developed a U-Net model, which was treated as the baseline model for this research, to show how the eye-gaze data can be used in multi-modal training for explainability improvement and disease classification. To compare the classification performances among this research's three experiment sets and the baseline model, the 95% confidence intervals (CI) of the area under the receiver operating characteristic curve (AUC) were measured. The best method achieved an AUC of 0.913 with a 95% CI of [0.860, 0.966]. \"Pneumonia\" and \"CHF\" classes, which the baseline model struggled the most to classify, had the greatest improvements, resulting in AUCs of 0.859 with a 95% CI of [0.732, 0.957] and 0.962 with a 95% CI of [0.933, 0.989], respectively. The decoder of the U-Net for the best-performing proposed method generated heatmaps that highlight the determining image parts in model classifications. These predicted heatmaps, which can be used for the explainability of the model, also improved to align well with the radiologist's eye-gaze data. Hence, this work showed that incorporating heatmap generators and eye-gaze information into training can simultaneously improve disease classification and provide explainable visuals that align well with how the radiologist viewed the chest radiographs when making diagnosis.</p>","PeriodicalId":73101,"journal":{"name":"Frontiers in radiology","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10365129/pdf/","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in radiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fradi.2022.991683","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

As deep learning is widely used in the radiology field, the explainability of Artificial Intelligence (AI) models is becoming increasingly essential to gain clinicians' trust when using the models for diagnosis. In this research, three experiment sets were conducted with a U-Net architecture to improve the disease classification performance while enhancing the heatmaps corresponding to the model's focus through incorporating heatmap generators during training. All experiments used the dataset that contained chest radiographs, associated labels from one of the three conditions ["normal", "congestive heart failure (CHF)", and "pneumonia"], and numerical information regarding a radiologist's eye-gaze coordinates on the images. The paper that introduced this dataset developed a U-Net model, which was treated as the baseline model for this research, to show how the eye-gaze data can be used in multi-modal training for explainability improvement and disease classification. To compare the classification performances among this research's three experiment sets and the baseline model, the 95% confidence intervals (CI) of the area under the receiver operating characteristic curve (AUC) were measured. The best method achieved an AUC of 0.913 with a 95% CI of [0.860, 0.966]. "Pneumonia" and "CHF" classes, which the baseline model struggled the most to classify, had the greatest improvements, resulting in AUCs of 0.859 with a 95% CI of [0.732, 0.957] and 0.962 with a 95% CI of [0.933, 0.989], respectively. The decoder of the U-Net for the best-performing proposed method generated heatmaps that highlight the determining image parts in model classifications. These predicted heatmaps, which can be used for the explainability of the model, also improved to align well with the radiologist's eye-gaze data. Hence, this work showed that incorporating heatmap generators and eye-gaze information into training can simultaneously improve disease classification and provide explainable visuals that align well with how the radiologist viewed the chest radiographs when making diagnosis.

Abstract Image

Abstract Image

Abstract Image

利用热图生成器改进放射学中疾病分类性能和深度学习模型的可解释性。
随着深度学习在放射学领域的广泛应用,人工智能(AI)模型的可解释性对于在使用模型进行诊断时获得临床医生的信任变得越来越重要。在本研究中,采用U-Net架构进行了三个实验集,以提高疾病分类性能,同时通过在训练过程中加入热图生成器来增强模型焦点对应的热图。所有实验使用的数据集包含胸片、三种情况(“正常”、“充血性心力衰竭”和“肺炎”)之一的相关标签,以及放射科医生在图像上的眼睛注视坐标的数字信息。介绍该数据集的论文开发了一个U-Net模型,该模型被视为本研究的基线模型,以展示如何将眼球注视数据用于多模式训练,以提高可解释性和疾病分类。为了比较本研究三个实验集与基线模型的分类性能,测量受试者工作特征曲线下面积(AUC)的95%置信区间(CI)。最佳方法的AUC为0.913,95% CI为[0.860,0.966]。基线模型最难分类的“肺炎”和“瑞士法郎”类别改善最大,auc分别为0.859,95% CI为[0.732,0.957]和0.962,95% CI为[0.933,0.989]。U-Net的解码器为表现最好的提出的方法生成热图,在模型分类中突出显示确定的图像部分。这些预测的热图可以用于模型的可解释性,也可以改进以与放射科医生的眼睛注视数据保持一致。因此,这项工作表明,将热图生成器和眼睛注视信息纳入培训可以同时改善疾病分类,并提供可解释的视觉效果,这些视觉效果与放射科医生在诊断时如何查看胸部x线片非常吻合。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
1.20
自引率
0.00%
发文量
0
文献相关原料
公司名称 产品信息 采购帮参考价格
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信