Privacy and Fairness in Machine Learning: A Survey

Sina Shaham;Arash Hajisafi;Minh K. Quan;Dinh C. Nguyen;Bhaskar Krishnamachari;Charith Peris;Gabriel Ghinita;Cyrus Shahabi;Pubudu N. Pathirana
{"title":"Privacy and Fairness in Machine Learning: A Survey","authors":"Sina Shaham;Arash Hajisafi;Minh K. Quan;Dinh C. Nguyen;Bhaskar Krishnamachari;Charith Peris;Gabriel Ghinita;Cyrus Shahabi;Pubudu N. Pathirana","doi":"10.1109/TAI.2025.3531326","DOIUrl":null,"url":null,"abstract":"Privacy and fairness are two crucial pillars of responsible artificial intelligence (AI) and trustworthy machine learning (ML). Each objective has been independently studied in the literature with the aim of reducing utility loss in achieving them. Despite the significant interest attracted from both academia and industry, there remains an immediate demand for more in-depth research to unravel how these two objectives can be simultaneously integrated into ML models. As opposed to well-accepted trade-offs, i.e., privacy-utility and fairness-utility, the interrelation between privacy and fairness is not well-understood. While some works suggest a trade-off between the two objective functions, there are others that demonstrate the alignment of these functions in certain scenarios. To fill this research gap, we provide a thorough review of privacy and fairness in ML, including supervised, unsupervised, semisupervised, and reinforcement learning. After examining and consolidating the literature on both objectives, we present a holistic survey on the impact of privacy on fairness, the impact of fairness on privacy, existing architectures, their interaction in application domains, and algorithms that aim to achieve both objectives while minimizing the utility sacrificed. Finally, we identify research challenges in achieving concurrently privacy and fairness in ML, particularly focusing on large language models.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 7","pages":"1706-1726"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10848485/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Privacy and fairness are two crucial pillars of responsible artificial intelligence (AI) and trustworthy machine learning (ML). Each objective has been independently studied in the literature with the aim of reducing utility loss in achieving them. Despite the significant interest attracted from both academia and industry, there remains an immediate demand for more in-depth research to unravel how these two objectives can be simultaneously integrated into ML models. As opposed to well-accepted trade-offs, i.e., privacy-utility and fairness-utility, the interrelation between privacy and fairness is not well-understood. While some works suggest a trade-off between the two objective functions, there are others that demonstrate the alignment of these functions in certain scenarios. To fill this research gap, we provide a thorough review of privacy and fairness in ML, including supervised, unsupervised, semisupervised, and reinforcement learning. After examining and consolidating the literature on both objectives, we present a holistic survey on the impact of privacy on fairness, the impact of fairness on privacy, existing architectures, their interaction in application domains, and algorithms that aim to achieve both objectives while minimizing the utility sacrificed. Finally, we identify research challenges in achieving concurrently privacy and fairness in ML, particularly focusing on large language models.
机器学习中的隐私和公平:一项调查
隐私和公平是负责任的人工智能(AI)和可信赖的机器学习(ML)的两个关键支柱。每个目标都在文献中进行了独立研究,目的是减少实现这些目标的效用损失。尽管学术界和工业界对此都很感兴趣,但仍然迫切需要进行更深入的研究,以揭示如何将这两个目标同时集成到ML模型中。与被广泛接受的权衡相反,即隐私-效用和公平-效用,隐私和公平之间的相互关系并没有得到很好的理解。虽然一些工作表明两个目标函数之间的权衡,但也有其他工作表明在某些情况下这些功能是一致的。为了填补这一研究空白,我们对机器学习中的隐私和公平性进行了全面的回顾,包括监督学习、无监督学习、半监督学习和强化学习。在研究和整合了关于这两个目标的文献之后,我们对隐私对公平的影响、公平对隐私的影响、现有架构、它们在应用领域中的相互作用以及旨在实现这两个目标同时最小化效用牺牲的算法进行了全面调查。最后,我们确定了在ML中同时实现隐私和公平的研究挑战,特别是关注大型语言模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
7.70
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信