Enabling Equal Opportunity in Logistic Regression Algorithm

S. Radovanović, M. Ivić
{"title":"Enabling Equal Opportunity in Logistic Regression Algorithm","authors":"S. Radovanović, M. Ivić","doi":"10.7595/management.fon.2021.0029","DOIUrl":null,"url":null,"abstract":"Research Question: This paper aims at adjusting the logistic regression algorithm to mitigate unwanted discrimination shown towards race, gender, etc. Motivation: Decades of research in the field of algorithm design have been dedicated to making a better prediction model. Many algorithms are designed and improved, which made them better than the judgments of people and even experts. However, in recent years it has been discovered that predictive models can make unwanted discrimination. Such unwanted discrimination in the predictive model can lead to legal consequences. In order to mitigate the problem of unwanted discrimination, we propose equal opportunity between privileged and discriminated groups in the logistic regression algorithm. Idea: Our idea is to add a regularization term in the goal function of the logistic regression. Therefore, our predictive model will solve both the social problem and the predictive problem. More specifically, our model will provide fair and accurate predictions. Data: The data used in this research present U.S. census data describing individuals using personal characteristics with a goal to provide a binary classification model for predicting if an individual has an annual salary above $50k. The dataset used is known for disparate impact regarding female individuals. In addition, we used the COMPAS dataset aimed at predicting recidivism. COMPAS is biased toward African-Americans. Tools: We developed a novel regularization technique for equal opportunity in the logistic regression algorithm. The proposed regularization is compared against classical logistic regression and fairness constraint logistic regression, using a ten-fold cross-validation. Findings: The results suggest that equal opportunity logistic regression manages to create a fair prediction model. More specifically, our model improved both disparate impact and equal opportunity compared to classical logistic regression, with a minor loss in prediction accuracy. Compared to the disparate impact constrained logistic regression, our approach has higher prediction accuracy and equal opportunity, while having a lower disparate impact. By inspecting the coefficients of our approach and classical logistic regression, one can see that proxy attribute coefficients are reduced to very low values. Contribution: The main contribution of this paper is in the methodological part. More specifically, we implemented an equal opportunity in the logistic regression algorithm.","PeriodicalId":429254,"journal":{"name":"Management:Journal of Sustainable Business and Management Solutions in Emerging Economies","volume":"110 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Management:Journal of Sustainable Business and Management Solutions in Emerging Economies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.7595/management.fon.2021.0029","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Research Question: This paper aims at adjusting the logistic regression algorithm to mitigate unwanted discrimination shown towards race, gender, etc. Motivation: Decades of research in the field of algorithm design have been dedicated to making a better prediction model. Many algorithms are designed and improved, which made them better than the judgments of people and even experts. However, in recent years it has been discovered that predictive models can make unwanted discrimination. Such unwanted discrimination in the predictive model can lead to legal consequences. In order to mitigate the problem of unwanted discrimination, we propose equal opportunity between privileged and discriminated groups in the logistic regression algorithm. Idea: Our idea is to add a regularization term in the goal function of the logistic regression. Therefore, our predictive model will solve both the social problem and the predictive problem. More specifically, our model will provide fair and accurate predictions. Data: The data used in this research present U.S. census data describing individuals using personal characteristics with a goal to provide a binary classification model for predicting if an individual has an annual salary above $50k. The dataset used is known for disparate impact regarding female individuals. In addition, we used the COMPAS dataset aimed at predicting recidivism. COMPAS is biased toward African-Americans. Tools: We developed a novel regularization technique for equal opportunity in the logistic regression algorithm. The proposed regularization is compared against classical logistic regression and fairness constraint logistic regression, using a ten-fold cross-validation. Findings: The results suggest that equal opportunity logistic regression manages to create a fair prediction model. More specifically, our model improved both disparate impact and equal opportunity compared to classical logistic regression, with a minor loss in prediction accuracy. Compared to the disparate impact constrained logistic regression, our approach has higher prediction accuracy and equal opportunity, while having a lower disparate impact. By inspecting the coefficients of our approach and classical logistic regression, one can see that proxy attribute coefficients are reduced to very low values. Contribution: The main contribution of this paper is in the methodological part. More specifically, we implemented an equal opportunity in the logistic regression algorithm.
在逻辑回归算法中实现机会均等
研究问题:本文旨在调整逻辑回归算法,以减轻对种族,性别等的不必要歧视。动机:几十年来,算法设计领域的研究一直致力于做出更好的预测模型。许多算法被设计和改进,使其优于人类甚至专家的判断。然而,近年来人们发现,预测模型可能会造成不必要的歧视。预测模型中这种不必要的歧视可能导致法律后果。为了减少不必要的歧视问题,我们在逻辑回归算法中提出特权群体和受歧视群体之间的机会均等。思路:我们的思路是在逻辑回归的目标函数中加入一个正则化项。因此,我们的预测模型将解决社会问题和预测问题。更具体地说,我们的模型将提供公平和准确的预测。数据:本研究中使用的数据是美国人口普查数据,描述了个人的个人特征,目的是提供一个二元分类模型,用于预测一个人的年薪是否超过5万美元。所使用的数据集以对女性个体的不同影响而闻名。此外,我们还使用了COMPAS数据集来预测再犯。COMPAS偏向非洲裔美国人。工具:我们在逻辑回归算法中开发了一种新的机会均等正则化技术。使用十倍交叉验证,将提出的正则化与经典逻辑回归和公平约束逻辑回归进行比较。结果表明,均等机会逻辑回归能够建立一个公平的预测模型。更具体地说,与经典逻辑回归相比,我们的模型改进了差异性影响和平等机会,预测精度略有下降。与异构影响约束逻辑回归相比,我们的方法具有更高的预测精度和均等的机会,同时具有更低的异构影响。通过检查我们的方法和经典逻辑回归的系数,可以看到代理属性系数被降低到非常低的值。贡献:本文的主要贡献在于方法论部分。更具体地说,我们在逻辑回归算法中实现了机会均等。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信