Stochastic Differentially Private and Fair Learning

Andrew Lowy, Devansh Gupta, Meisam Razaviyayn
{"title":"Stochastic Differentially Private and Fair Learning","authors":"Andrew Lowy, Devansh Gupta, Meisam Razaviyayn","doi":"10.48550/arXiv.2210.08781","DOIUrl":null,"url":null,"abstract":"Machine learning models are increasingly used in high-stakes decision-making systems. In such applications, a major concern is that these models sometimes discriminate against certain demographic groups such as individuals with certain race, gender, or age. Another major concern in these applications is the violation of the privacy of users. While fair learning algorithms have been developed to mitigate discrimination issues, these algorithms can still leak sensitive information, such as individuals' health or financial records. Utilizing the notion of differential privacy (DP), prior works aimed at developing learning algorithms that are both private and fair. However, existing algorithms for DP fair learning are either not guaranteed to converge or require full batch of data in each iteration of the algorithm to converge. In this paper, we provide the first stochastic differentially private algorithm for fair learning that is guaranteed to converge. Here, the term\"stochastic\"refers to the fact that our proposed algorithm converges even when minibatches of data are used at each iteration (i.e. stochastic optimization). Our framework is flexible enough to permit different fairness notions, including demographic parity and equalized odds. In addition, our algorithm can be applied to non-binary classification tasks with multiple (non-binary) sensitive attributes. As a byproduct of our convergence analysis, we provide the first utility guarantee for a DP algorithm for solving nonconvex-strongly concave min-max problems. Our numerical experiments show that the proposed algorithm consistently offers significant performance gains over the state-of-the-art baselines, and can be applied to larger scale problems with non-binary target/sensitive attributes.","PeriodicalId":189973,"journal":{"name":"International Conference on Learning Representations","volume":"26 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Learning Representations","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2210.08781","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Machine learning models are increasingly used in high-stakes decision-making systems. In such applications, a major concern is that these models sometimes discriminate against certain demographic groups such as individuals with certain race, gender, or age. Another major concern in these applications is the violation of the privacy of users. While fair learning algorithms have been developed to mitigate discrimination issues, these algorithms can still leak sensitive information, such as individuals' health or financial records. Utilizing the notion of differential privacy (DP), prior works aimed at developing learning algorithms that are both private and fair. However, existing algorithms for DP fair learning are either not guaranteed to converge or require full batch of data in each iteration of the algorithm to converge. In this paper, we provide the first stochastic differentially private algorithm for fair learning that is guaranteed to converge. Here, the term"stochastic"refers to the fact that our proposed algorithm converges even when minibatches of data are used at each iteration (i.e. stochastic optimization). Our framework is flexible enough to permit different fairness notions, including demographic parity and equalized odds. In addition, our algorithm can be applied to non-binary classification tasks with multiple (non-binary) sensitive attributes. As a byproduct of our convergence analysis, we provide the first utility guarantee for a DP algorithm for solving nonconvex-strongly concave min-max problems. Our numerical experiments show that the proposed algorithm consistently offers significant performance gains over the state-of-the-art baselines, and can be applied to larger scale problems with non-binary target/sensitive attributes.
随机差异、私有和公平学习
机器学习模型越来越多地用于高风险决策系统。在这样的应用程序中,一个主要的问题是这些模型有时会歧视某些人口统计群体,例如具有特定种族、性别或年龄的个人。这些应用程序的另一个主要问题是侵犯用户的隐私。虽然已经开发了公平学习算法来减轻歧视问题,但这些算法仍然可能泄露敏感信息,例如个人健康或财务记录。利用差分隐私(DP)的概念,先前的工作旨在开发既私密又公平的学习算法。然而,现有的DP公平学习算法要么不能保证收敛,要么每次迭代都需要整批数据才能收敛。在本文中,我们提供了第一个保证收敛的公平学习随机差分私有算法。在这里,术语“随机”指的是我们提出的算法即使在每次迭代中使用小批量数据(即随机优化)时也是收敛的。我们的框架足够灵活,可以允许不同的公平概念,包括人口均等和均等几率。此外,我们的算法可以应用于具有多个(非二进制)敏感属性的非二进制分类任务。作为我们收敛性分析的副产品,我们为求解非凸强凹最小-最大问题的DP算法提供了第一个效用保证。我们的数值实验表明,所提出的算法在最先进的基线上始终提供显著的性能提升,并且可以应用于具有非二进制目标/敏感属性的更大规模问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信