{"title":"Meta-XAI for Explaining the Explainer: Unveiling Image Features Driving Deep Learning Decisions","authors":"Simone Bianco","doi":"10.1109/TAI.2025.3529397","DOIUrl":null,"url":null,"abstract":"Deep learning has revolutionized computer vision by allowing neural networks to automatically learn features from data. However, the highly nonlinear nature of deep neural networks makes them difficult to interpret, leading to concerns about potential biases in critical applications. To address this, researchers have advocated for explainable artificial intelligence (XAI). Many XAI techniques have been proposed but all of them only highlight image regions influencing model decisions, lacking any further explanations. In this article, we propose a posthoc model-agnostic meta-XAI method that explains why specific image regions are used for decisions. The article presents the experimental setup and results, discussing the perturbations used for explanations in color, frequency, shape, shading, and texture. The explanation is given in terms of human-interpretable image features, e.g., color, shape, shading, and texture both as perturbation plots and as visual summary through the use of the newly introduced normalized area under the curve score. The experimental results confirm the previous findings that vision deep learning models are biased toward texture, but also highlight the importance of color, frequency content, and perceptually salient structures in the final decision.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 7","pages":"1859-1869"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10839578/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Deep learning has revolutionized computer vision by allowing neural networks to automatically learn features from data. However, the highly nonlinear nature of deep neural networks makes them difficult to interpret, leading to concerns about potential biases in critical applications. To address this, researchers have advocated for explainable artificial intelligence (XAI). Many XAI techniques have been proposed but all of them only highlight image regions influencing model decisions, lacking any further explanations. In this article, we propose a posthoc model-agnostic meta-XAI method that explains why specific image regions are used for decisions. The article presents the experimental setup and results, discussing the perturbations used for explanations in color, frequency, shape, shading, and texture. The explanation is given in terms of human-interpretable image features, e.g., color, shape, shading, and texture both as perturbation plots and as visual summary through the use of the newly introduced normalized area under the curve score. The experimental results confirm the previous findings that vision deep learning models are biased toward texture, but also highlight the importance of color, frequency content, and perceptually salient structures in the final decision.