Souichi Oka , Takuma Yamazaki , Yoshiyasu Takefuji
{"title":"Pitfalls of XAI interpretation in environmental modeling: A warning on model bias in air quality data analysis","authors":"Souichi Oka , Takuma Yamazaki , Yoshiyasu Takefuji","doi":"10.1016/j.envsoft.2025.106700","DOIUrl":null,"url":null,"abstract":"<div><div>Jung et al. (2025) achieved high predictive accuracy in interpolating missing ozone data using graph machine learning (ML) and conducted feature importance analysis with explainable AI (XAI). This correspondence acknowledges their significant contribution but discusses the limitations and biases inherent in ML models and XAI methods (e.g., Random Forest/Bootstrap Test, SHapley Additive exPlanations (SHAP)) and their impact on the reliability of derived feature importance. High predictive accuracy does not necessarily guarantee trustworthy interpretation of feature relevance, as evidenced by inconsistent importance rankings across models and XAI techniques. To enhance interpretability and scientific reliability, we advocate a validation strategy integrating ML with rigorous statistical analysis. It combines model-driven insights with statistical measures such as Spearman's rho and Kendall's tau, and information-theoretic metrics like Mutual Information and Total Correlation to capture complex, non-linear dependencies. Such integration improves the robustness of feature importance assessments and supports more reliable interpretations in environmental modeling.</div></div>","PeriodicalId":310,"journal":{"name":"Environmental Modelling & Software","volume":"194 ","pages":"Article 106700"},"PeriodicalIF":4.6000,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Environmental Modelling & Software","FirstCategoryId":"93","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1364815225003846","RegionNum":2,"RegionCategory":"环境科学与生态学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
Jung et al. (2025) achieved high predictive accuracy in interpolating missing ozone data using graph machine learning (ML) and conducted feature importance analysis with explainable AI (XAI). This correspondence acknowledges their significant contribution but discusses the limitations and biases inherent in ML models and XAI methods (e.g., Random Forest/Bootstrap Test, SHapley Additive exPlanations (SHAP)) and their impact on the reliability of derived feature importance. High predictive accuracy does not necessarily guarantee trustworthy interpretation of feature relevance, as evidenced by inconsistent importance rankings across models and XAI techniques. To enhance interpretability and scientific reliability, we advocate a validation strategy integrating ML with rigorous statistical analysis. It combines model-driven insights with statistical measures such as Spearman's rho and Kendall's tau, and information-theoretic metrics like Mutual Information and Total Correlation to capture complex, non-linear dependencies. Such integration improves the robustness of feature importance assessments and supports more reliable interpretations in environmental modeling.
期刊介绍:
Environmental Modelling & Software publishes contributions, in the form of research articles, reviews and short communications, on recent advances in environmental modelling and/or software. The aim is to improve our capacity to represent, understand, predict or manage the behaviour of environmental systems at all practical scales, and to communicate those improvements to a wide scientific and professional audience.