Clarity in complexity: how aggregating explanations resolves the disagreement problem

IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Oana Mitruț, Gabriela Moise, Alin Moldoveanu, Florica Moldoveanu, Marius Leordeanu, Livia Petrescu
{"title":"Clarity in complexity: how aggregating explanations resolves the disagreement problem","authors":"Oana Mitruț,&nbsp;Gabriela Moise,&nbsp;Alin Moldoveanu,&nbsp;Florica Moldoveanu,&nbsp;Marius Leordeanu,&nbsp;Livia Petrescu","doi":"10.1007/s10462-024-10952-7","DOIUrl":null,"url":null,"abstract":"<div><p>The Rashômon Effect, applied in Explainable Machine Learning, refers to the disagreement between the explanations provided by various attribution explainers and to the dissimilarity across multiple explanations generated by a particular explainer for a single instance from the dataset (differences between feature importances and their associated signs and ranks), an undesirable outcome especially in sensitive domains such as healthcare or finance. We propose a method inspired from textual-case based reasoning for aligning explanations from various explainers in order to resolve the disagreement and dissimilarity problems. We iteratively generated a number of 100 explanations for each instance from six popular datasets, using three prevalent feature attribution explainers: LIME, Anchors and SHAP (with the variations Tree SHAP and Kernel SHAP) and consequently applied a global cluster-based aggregation strategy that quantifies alignment and reveals similarities and associations between explanations. We evaluated our method by weighting the <span>\\(\\:k\\)</span>-NN algorithm with agreed feature overlap explanation weights and compared it to a non-weighted <span>\\(\\:k\\)</span>-NN predictor, having as task binary classification. Also, we compared the results of the weighted <span>\\(\\:k\\)</span>-NN algorithm using aggregated feature overlap explanation weights to the weighted <span>\\(\\:k\\)</span>-NN algorithm using weights produced by a single explanation method (either LIME, SHAP or Anchors). Our global alignment method benefited the most from a hybridization with feature importance scores (information gain), that was essential for acquiring a more accurate estimate of disagreement, for enabling explainers to reach a consensus across multiple explanations and for supporting effective model learning through improved classification performance.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"57 12","pages":""},"PeriodicalIF":10.7000,"publicationDate":"2024-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10952-7.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Intelligence Review","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10462-024-10952-7","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

The Rashômon Effect, applied in Explainable Machine Learning, refers to the disagreement between the explanations provided by various attribution explainers and to the dissimilarity across multiple explanations generated by a particular explainer for a single instance from the dataset (differences between feature importances and their associated signs and ranks), an undesirable outcome especially in sensitive domains such as healthcare or finance. We propose a method inspired from textual-case based reasoning for aligning explanations from various explainers in order to resolve the disagreement and dissimilarity problems. We iteratively generated a number of 100 explanations for each instance from six popular datasets, using three prevalent feature attribution explainers: LIME, Anchors and SHAP (with the variations Tree SHAP and Kernel SHAP) and consequently applied a global cluster-based aggregation strategy that quantifies alignment and reveals similarities and associations between explanations. We evaluated our method by weighting the \(\:k\)-NN algorithm with agreed feature overlap explanation weights and compared it to a non-weighted \(\:k\)-NN predictor, having as task binary classification. Also, we compared the results of the weighted \(\:k\)-NN algorithm using aggregated feature overlap explanation weights to the weighted \(\:k\)-NN algorithm using weights produced by a single explanation method (either LIME, SHAP or Anchors). Our global alignment method benefited the most from a hybridization with feature importance scores (information gain), that was essential for acquiring a more accurate estimate of disagreement, for enabling explainers to reach a consensus across multiple explanations and for supporting effective model learning through improved classification performance.

复杂中的清晰:汇总解释如何解决分歧问题
可解释机器学习(Explainable Machine Learning)中应用的 "罗生门效应"(Rashômon Effect)是指不同归因解释者提供的解释之间存在分歧,以及特定解释者针对数据集中的单个实例生成的多个解释之间存在不相似性(特征导入量及其相关符号和等级之间存在差异)。我们从基于文本案例的推理中获得灵感,提出了一种方法来对齐来自不同解释者的解释,以解决分歧和差异问题。我们使用三种流行的特征归因解释器,从六个流行数据集的每个实例中反复生成了 100 个解释:我们使用 LIME、Anchors 和 SHAP(包括 Tree SHAP 和 Kernel SHAP 变体)这三种流行的特征归因解释器为六个流行数据集的每个实例生成了 100 个解释,并因此应用了一种基于聚类的全局聚合策略,该策略可量化对齐情况并揭示解释之间的相似性和关联性。我们通过使用商定的特征重叠解释权重对(\:k\)-NN 算法进行加权来评估我们的方法,并将其与任务为二元分类的非加权(\:k\)-NN 预测器进行比较。此外,我们还比较了使用聚合特征重叠解释权重的加权(\:k\)-NN 算法和使用单一解释方法(LIME、SHAP 或 Anchors)产生的权重的加权(\:k\)-NN 算法的结果。我们的全局配准方法从与特征重要性得分(信息增益)的混合中获益最大,这对于获得更准确的分歧估计、使解释者在多个解释中达成共识以及通过提高分类性能支持有效的模型学习至关重要。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Artificial Intelligence Review
Artificial Intelligence Review 工程技术-计算机:人工智能
CiteScore
22.00
自引率
3.30%
发文量
194
审稿时长
5.3 months
期刊介绍: Artificial Intelligence Review, a fully open access journal, publishes cutting-edge research in artificial intelligence and cognitive science. It features critical evaluations of applications, techniques, and algorithms, providing a platform for both researchers and application developers. The journal includes refereed survey and tutorial articles, along with reviews and commentary on significant developments in the field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信