The Crucial Role of Sensitive Attributes in Fair Classification

M. Haeri, K. Zweig
{"title":"The Crucial Role of Sensitive Attributes in Fair Classification","authors":"M. Haeri, K. Zweig","doi":"10.1109/SSCI47803.2020.9308585","DOIUrl":null,"url":null,"abstract":"In many countries, it is illegal to make certain decisions based on sensitive attributes such as gender or race. This is because historically, sensitive attributes of individuals were exploited to abuse the rights of individuals, leading to unfair decisions. This view is extended to algorithmic decision-making systems (ADMs) where similar to humans, ADMs should not use sensitive attributes for input. We reject the extension of law from humans to machines, since contrary to humans, algorithms are explicit in their decisions, and the fairness of their decision can be studied independently of their input. The main purpose of this paper is to study and discuss the importance of using sensitive attributes in fair classification systems. Specifically, we suggest two statistical tests on the training dataset, to evaluate whether using sensitive attributes may have an impact on the quality and fairness of prospective classification algorithms. These statistical tests compare the distribution and data complexity of the training dataset between groups identified by the same value for sensitive attributes (e.g., men vs. women). We evaluated our fairness tests on several datasets. It was shown that, the removal of sensitive attributes may result in the decrease of the fairness of ADMs. The results were confirmed by designing and implementing simple classifiers on each dataset (with and without the sensitive attributes). Therefore, the use of sensitive attributes must be evaluated per dataset and algorithm, and ignoring them blindly may result in unfair ADMs.","PeriodicalId":413489,"journal":{"name":"2020 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI47803.2020.9308585","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

In many countries, it is illegal to make certain decisions based on sensitive attributes such as gender or race. This is because historically, sensitive attributes of individuals were exploited to abuse the rights of individuals, leading to unfair decisions. This view is extended to algorithmic decision-making systems (ADMs) where similar to humans, ADMs should not use sensitive attributes for input. We reject the extension of law from humans to machines, since contrary to humans, algorithms are explicit in their decisions, and the fairness of their decision can be studied independently of their input. The main purpose of this paper is to study and discuss the importance of using sensitive attributes in fair classification systems. Specifically, we suggest two statistical tests on the training dataset, to evaluate whether using sensitive attributes may have an impact on the quality and fairness of prospective classification algorithms. These statistical tests compare the distribution and data complexity of the training dataset between groups identified by the same value for sensitive attributes (e.g., men vs. women). We evaluated our fairness tests on several datasets. It was shown that, the removal of sensitive attributes may result in the decrease of the fairness of ADMs. The results were confirmed by designing and implementing simple classifiers on each dataset (with and without the sensitive attributes). Therefore, the use of sensitive attributes must be evaluated per dataset and algorithm, and ignoring them blindly may result in unfair ADMs.
敏感属性在公平分类中的关键作用
在许多国家,基于性别或种族等敏感属性做出某些决定是非法的。这是因为在历史上,个人的敏感属性被利用来滥用个人的权利,导致不公平的决定。这种观点被扩展到算法决策系统(ADMs)中,与人类相似,ADMs不应该使用敏感属性作为输入。我们拒绝将法律从人类延伸到机器,因为与人类相反,算法在它们的决策中是明确的,它们的决策的公平性可以独立于它们的输入来研究。本文的主要目的是研究和讨论在公平分类系统中使用敏感属性的重要性。具体来说,我们建议在训练数据集上进行两个统计测试,以评估使用敏感属性是否会对未来分类算法的质量和公平性产生影响。这些统计测试在敏感属性(例如,男性与女性)的相同值识别的组之间比较训练数据集的分布和数据复杂性。我们在几个数据集上评估了公平性测试。结果表明,去除敏感属性会导致adm的公平性降低。通过在每个数据集上设计和实现简单的分类器(带和不带敏感属性)来验证结果。因此,敏感属性的使用必须对每个数据集和算法进行评估,盲目忽略它们可能会导致不公平的adm。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信