PreCoF: counterfactual explanations for fairness.

IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Sofie Goethals, David Martens, Toon Calders
{"title":"<i>PreCoF</i>: counterfactual explanations for fairness.","authors":"Sofie Goethals,&nbsp;David Martens,&nbsp;Toon Calders","doi":"10.1007/s10994-023-06319-8","DOIUrl":null,"url":null,"abstract":"<p><p>This paper studies how counterfactual explanations can be used to assess the fairness of a model. Using machine learning for high-stakes decisions is a threat to fairness as these models can amplify bias present in the dataset, and there is no consensus on a universal metric to detect this. The appropriate metric and method to tackle the bias in a dataset will be case-dependent, and it requires insight into the nature of the bias first. We aim to provide this insight by integrating explainable AI (XAI) research with the fairness domain. More specifically, apart from being able to use (Predictive) Counterfactual Explanations to detect <i>explicit bias</i> when the model is directly using the sensitive attribute, we show that it can also be used to detect <i>implicit bias</i> when the model does not use the sensitive attribute directly but does use other correlated attributes leading to a substantial disadvantage for a protected group. We call this metric <i>PreCoF</i>, or Predictive Counterfactual Fairness. Our experimental results show that our metric succeeds in detecting occurrences of <i>implicit bias</i> in the model by assessing which attributes are more present in the explanations of the protected group compared to the unprotected group. These results could help policymakers decide on whether this discrimination is <i>justified</i> or not.</p>","PeriodicalId":49900,"journal":{"name":"Machine Learning","volume":" ","pages":"1-32"},"PeriodicalIF":4.3000,"publicationDate":"2023-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10047477/pdf/","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine Learning","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s10994-023-06319-8","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 5

Abstract

This paper studies how counterfactual explanations can be used to assess the fairness of a model. Using machine learning for high-stakes decisions is a threat to fairness as these models can amplify bias present in the dataset, and there is no consensus on a universal metric to detect this. The appropriate metric and method to tackle the bias in a dataset will be case-dependent, and it requires insight into the nature of the bias first. We aim to provide this insight by integrating explainable AI (XAI) research with the fairness domain. More specifically, apart from being able to use (Predictive) Counterfactual Explanations to detect explicit bias when the model is directly using the sensitive attribute, we show that it can also be used to detect implicit bias when the model does not use the sensitive attribute directly but does use other correlated attributes leading to a substantial disadvantage for a protected group. We call this metric PreCoF, or Predictive Counterfactual Fairness. Our experimental results show that our metric succeeds in detecting occurrences of implicit bias in the model by assessing which attributes are more present in the explanations of the protected group compared to the unprotected group. These results could help policymakers decide on whether this discrimination is justified or not.

Abstract Image

Abstract Image

Abstract Image

PreCoF:对公平的反事实解释。
本文研究如何使用反事实解释来评估模型的公平性。将机器学习用于高风险决策是对公平性的威胁,因为这些模型可能会放大数据集中存在的偏见,而且对于检测这一点的通用指标还没有达成共识。解决数据集中偏差的适当指标和方法将取决于具体情况,需要首先深入了解偏差的性质。我们的目标是通过将可解释人工智能(XAI)研究与公平领域相结合来提供这一见解。更具体地说,当模型直接使用敏感属性时,除了能够使用(预测)反事实解释来检测显式偏见外,我们还表明,当模型不直接使用敏感内容,但使用了其他相关属性,导致受保护群体处于严重劣势时,它也可以用于检测隐式偏见。我们将这种度量称为PreCoF,或预测反事实公平性。我们的实验结果表明,我们的度量通过评估与未保护组相比,受保护组的解释中哪些属性更具代表性,成功地检测到了模型中隐性偏见的发生。这些结果可以帮助决策者决定这种歧视是否合理。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Machine Learning
Machine Learning 工程技术-计算机:人工智能
CiteScore
11.00
自引率
2.70%
发文量
162
审稿时长
3 months
期刊介绍: Machine Learning serves as a global platform dedicated to computational approaches in learning. The journal reports substantial findings on diverse learning methods applied to various problems, offering support through empirical studies, theoretical analysis, or connections to psychological phenomena. It demonstrates the application of learning methods to solve significant problems and aims to enhance the conduct of machine learning research with a focus on verifiable and replicable evidence in published papers.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信