Maria Carla Calzarossa , Paolo Giudici , Rasha Zieni
{"title":"集成机器学习解释的鲁棒性如何?","authors":"Maria Carla Calzarossa , Paolo Giudici , Rasha Zieni","doi":"10.1016/j.neucom.2025.129686","DOIUrl":null,"url":null,"abstract":"<div><div>To date, several explainable AI methods are available. The variability of the resulting explanations can be high, especially when many input features are considered. This lack of robustness may limit their usability. In this paper we try to fill this gap, by contributing a methodology that: i) is able to measure the robustness of a given set of explanations; ii) suggests how to improve robustness, by tuning the model parameters. Without loss of generality, we exemplify our proposal for ensemble tree models, which typically reach a high predictive performance in classification problems. We consider a toy case study with artificially generated data as well as two real case studies whose application domain is cybersecurity and more precisely the models used for detecting phishing websites.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"630 ","pages":"Article 129686"},"PeriodicalIF":6.5000,"publicationDate":"2025-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"How robust are ensemble machine learning explanations?\",\"authors\":\"Maria Carla Calzarossa , Paolo Giudici , Rasha Zieni\",\"doi\":\"10.1016/j.neucom.2025.129686\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>To date, several explainable AI methods are available. The variability of the resulting explanations can be high, especially when many input features are considered. This lack of robustness may limit their usability. In this paper we try to fill this gap, by contributing a methodology that: i) is able to measure the robustness of a given set of explanations; ii) suggests how to improve robustness, by tuning the model parameters. Without loss of generality, we exemplify our proposal for ensemble tree models, which typically reach a high predictive performance in classification problems. We consider a toy case study with artificially generated data as well as two real case studies whose application domain is cybersecurity and more precisely the models used for detecting phishing websites.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"630 \",\"pages\":\"Article 129686\"},\"PeriodicalIF\":6.5000,\"publicationDate\":\"2025-02-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225003583\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225003583","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
How robust are ensemble machine learning explanations?
To date, several explainable AI methods are available. The variability of the resulting explanations can be high, especially when many input features are considered. This lack of robustness may limit their usability. In this paper we try to fill this gap, by contributing a methodology that: i) is able to measure the robustness of a given set of explanations; ii) suggests how to improve robustness, by tuning the model parameters. Without loss of generality, we exemplify our proposal for ensemble tree models, which typically reach a high predictive performance in classification problems. We consider a toy case study with artificially generated data as well as two real case studies whose application domain is cybersecurity and more precisely the models used for detecting phishing websites.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.