Ahmad Chaddad , Yuchen Jiang , Tareef S. Daqqaq , Reem Kateb
{"title":"EAMAPG: Explainable Adversarial Model Analysis via Projected Gradient Descent","authors":"Ahmad Chaddad , Yuchen Jiang , Tareef S. Daqqaq , Reem Kateb","doi":"10.1016/j.compbiomed.2025.109788","DOIUrl":null,"url":null,"abstract":"<div><div>Despite the outstanding performance of deep learning (DL) models, their interpretability remains a challenging topic. In this study, we address the transparency of DL models in medical image analysis by introducing a novel interpretability method using projected gradient descent (PGD) to generate adversarial examples. We use adversarial generation to analyze images. By introducing perturbations that cause misclassification, we identify key features influencing the model decisions. This method is tested on Brain Tumor, Eye Disease, and COVID-19 datasets using six common convolutional neural networks (CNN) models. We selected the top-performing models for interpretability analysis. DenseNet121 achieved an AUC of 1.00 on Brain Tumor; InceptionV3, 0.99 on Eye Disease; and ResNet101, 1.00 on COVID-19. To test their robustness, we performed an adversarial attack. The p-values from t-tests comparing original and adversarial loss distributions were all <span><math><mo><</mo></math></span> 0.05. This indicates that the adversarial perturbations significantly increased the loss, confirming successful adversarial generation. Our approach offers a distinct solution to bridge the gap between the capabilities of artificial intelligence and its practical use in clinical settings, providing a more intuitive understanding for radiologists. Our code is available at <span><span>https://anonymous.4open.science/r/EAMAPG</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":10578,"journal":{"name":"Computers in biology and medicine","volume":"188 ","pages":"Article 109788"},"PeriodicalIF":7.0000,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers in biology and medicine","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0010482525001386","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"BIOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Despite the outstanding performance of deep learning (DL) models, their interpretability remains a challenging topic. In this study, we address the transparency of DL models in medical image analysis by introducing a novel interpretability method using projected gradient descent (PGD) to generate adversarial examples. We use adversarial generation to analyze images. By introducing perturbations that cause misclassification, we identify key features influencing the model decisions. This method is tested on Brain Tumor, Eye Disease, and COVID-19 datasets using six common convolutional neural networks (CNN) models. We selected the top-performing models for interpretability analysis. DenseNet121 achieved an AUC of 1.00 on Brain Tumor; InceptionV3, 0.99 on Eye Disease; and ResNet101, 1.00 on COVID-19. To test their robustness, we performed an adversarial attack. The p-values from t-tests comparing original and adversarial loss distributions were all 0.05. This indicates that the adversarial perturbations significantly increased the loss, confirming successful adversarial generation. Our approach offers a distinct solution to bridge the gap between the capabilities of artificial intelligence and its practical use in clinical settings, providing a more intuitive understanding for radiologists. Our code is available at https://anonymous.4open.science/r/EAMAPG.
期刊介绍:
Computers in Biology and Medicine is an international forum for sharing groundbreaking advancements in the use of computers in bioscience and medicine. This journal serves as a medium for communicating essential research, instruction, ideas, and information regarding the rapidly evolving field of computer applications in these domains. By encouraging the exchange of knowledge, we aim to facilitate progress and innovation in the utilization of computers in biology and medicine.