Interpretable surrogate models to approximate the predictions of convolutional neural networks in glaucoma diagnosis

IF 6.3 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Jose Sigut, Francisco José Fumero Batista, Rafael Arnay, José Estévez, Tinguaro Díaz-Alemán
{"title":"Interpretable surrogate models to approximate the predictions of convolutional neural networks in glaucoma diagnosis","authors":"Jose Sigut, Francisco José Fumero Batista, Rafael Arnay, José Estévez, Tinguaro Díaz-Alemán","doi":"10.1088/2632-2153/ad0798","DOIUrl":null,"url":null,"abstract":"Abstract Background and objective
Deep learning systems, especially in critical fields like medicine, suffer from a significant drawback - their black box nature, which lacks mechanisms for explaining or interpreting their decisions. In this regard, our research aims to evaluate the use of surrogate models for interpreting convolutional neural network decisions in glaucoma diagnosis. Our approach is novel in that we approximate the original model with an interpretable one and also change the input features, replacing pixels with tabular geometric features of the optic disc, cup, and neuroretinal rim.

Method
We trained convolutional neural networks with two types of images: original images of the optic nerve head and simplified images showing only the disc and cup contours on a uniform background. Decision trees were used as surrogate models due to their simplicity and visualization properties, while saliency maps were calculated for some images for comparison.

Results
The experiments carried out with 1271 images of healthy subjects and 721 images of glaucomatous eyes demonstrate that decision trees can closely approximate the predictions of neural networks trained on simplified contour images, with R-squared values near 0.9 for VGG19, Resnet50, InceptionV3 and Xception architectures. Saliency maps proved difficult to interpret and showed inconsistent results across architectures, in contrast to the decision trees. Additionally, some decision trees trained as surrogate models outperformed a decision tree trained on the actual outcomes without surrogation.

Conclusions
Decision trees may be a more interpretable alternative to saliency methods. Moreover, the fact that we matched the performance of a decision tree without surrogation to that obtained by decision trees using knowledge distillation from neural networks is a great advantage since decision trees are inherently interpretable. Therefore, based on our findings, we think this approach would be the most recommendable choice for specialists as a diagnostic tool.","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"24 3","pages":"0"},"PeriodicalIF":6.3000,"publicationDate":"2023-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine Learning Science and Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/2632-2153/ad0798","RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Abstract Background and objective
Deep learning systems, especially in critical fields like medicine, suffer from a significant drawback - their black box nature, which lacks mechanisms for explaining or interpreting their decisions. In this regard, our research aims to evaluate the use of surrogate models for interpreting convolutional neural network decisions in glaucoma diagnosis. Our approach is novel in that we approximate the original model with an interpretable one and also change the input features, replacing pixels with tabular geometric features of the optic disc, cup, and neuroretinal rim.

Method
We trained convolutional neural networks with two types of images: original images of the optic nerve head and simplified images showing only the disc and cup contours on a uniform background. Decision trees were used as surrogate models due to their simplicity and visualization properties, while saliency maps were calculated for some images for comparison.

Results
The experiments carried out with 1271 images of healthy subjects and 721 images of glaucomatous eyes demonstrate that decision trees can closely approximate the predictions of neural networks trained on simplified contour images, with R-squared values near 0.9 for VGG19, Resnet50, InceptionV3 and Xception architectures. Saliency maps proved difficult to interpret and showed inconsistent results across architectures, in contrast to the decision trees. Additionally, some decision trees trained as surrogate models outperformed a decision tree trained on the actual outcomes without surrogation.

Conclusions
Decision trees may be a more interpretable alternative to saliency methods. Moreover, the fact that we matched the performance of a decision tree without surrogation to that obtained by decision trees using knowledge distillation from neural networks is a great advantage since decision trees are inherently interpretable. Therefore, based on our findings, we think this approach would be the most recommendable choice for specialists as a diagnostic tool.
可解释的替代模型来近似卷积神经网络在青光眼诊断中的预测
深度学习系统,特别是在像医学这样的关键领域,有一个明显的缺点——它们的黑箱性质,缺乏解释或解释它们的决定的机制。在这方面,我们的研究旨在评估替代模型在青光眼诊断中解释卷积神经网络决策的使用。我们的方法是新颖的,因为我们用一个可解释的模型近似原始模型,并且还改变了输入特征,用视盘、杯和神经视网膜边缘的表格几何特征替换像素。方法我们用两种类型的图像训练卷积神经网络:视神经头的原始图像和在统一背景上仅显示盘和杯轮廓的简化图像。由于决策树的简单性和可视化特性,我们使用决策树作为替代模型,同时对一些图像计算显著性图进行比较。结果对1271张健康受试者图像和721张青光眼图像进行的实验表明,决策树可以非常接近在简化轮廓图像上训练的神经网络的预测,VGG19、Resnet50、Resnet50的r平方值接近0.9。InceptionV3和Xception架构。与决策树相比,显著性图被证明很难解释,并且在体系结构中显示不一致的结果。此外,一些作为替代模型训练的决策树比没有替代的实际结果训练的决策树表现得更好。结论:决策树可能是显著性方法的更可解释的替代方法。此外,由于决策树具有固有的可解释性,因此我们将不需要替代的决策树的性能与使用神经网络知识蒸馏的决策树的性能相匹配是一个很大的优势。因此,根据我们的研究结果,我们认为这种方法将是专家最推荐的诊断工具选择。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Machine Learning Science and Technology
Machine Learning Science and Technology Computer Science-Artificial Intelligence
CiteScore
9.10
自引率
4.40%
发文量
86
审稿时长
5 weeks
期刊介绍: Machine Learning Science and Technology is a multidisciplinary open access journal that bridges the application of machine learning across the sciences with advances in machine learning methods and theory as motivated by physical insights. Specifically, articles must fall into one of the following categories: advance the state of machine learning-driven applications in the sciences or make conceptual, methodological or theoretical advances in machine learning with applications to, inspiration from, or motivated by scientific problems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信