FairST: A novel approach for machine learning bias repair through latent sensitive attribute translation

IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS
Carmen Meinson , Max Hort , Federica Sarro
{"title":"FairST: A novel approach for machine learning bias repair through latent sensitive attribute translation","authors":"Carmen Meinson ,&nbsp;Max Hort ,&nbsp;Federica Sarro","doi":"10.1016/j.infsof.2025.107900","DOIUrl":null,"url":null,"abstract":"<div><h3>Context:</h3><div>As Machine Learning (ML) models are increasingly used in critical decision-making software, concerns have been raised about these systems perpetuating or exacerbating existing historical biases. Consequently, there has been a growing research interest in developing methods to test for fairness and repair biases in ML software, particularly for legally protected attributes like gender, age, race.</div></div><div><h3>Objectives:</h3><div>In this work, we set out to repair bias for both single and multiple protected attributes (a.k.a. intersectional fairness) of pre-trained machine learning models.</div></div><div><h3>Methods:</h3><div>We propose a novel model- and task-agnostic debiasing method, Fair Subgroup Translation (FairST), based on fair representation learning via auto-encoders. To the best of our knowledge, this is the first approach based on the principle of Fair Representation Learning devised for post-processing bias repair.</div></div><div><h3>Results:</h3><div>We empirically evaluate the effectiveness of using FairST to repair a pre-trained Neural Network model used for seven classification tasks involving both single and multiple protected attributes, and benchmark its performance with state-of-the-art fairness repair methods (i.e., Learning Fair Representations, Reweighing, FairBalance and FairMask). We also investigate if the effectiveness of FairST varies when using it to repair bias of other popular ML models (namely Logistic Regression, Support Vector Machine, Gaussian Naive Bayes, Decision Tree and Random Forest).</div></div><div><h3>Conclusion:</h3><div>The results demonstrate that FairST consistently achieves superior single and intersectional fairness with respect to all benchmarking methods for all classification tasks considered in our empirical study. This supports the potential of using FairST for ML bias repair, and opens up a rich agenda of future work including its application to repair bias arising in tasks of a different nature such as multi-class or image-based problems.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"189 ","pages":"Article 107900"},"PeriodicalIF":4.3000,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information and Software Technology","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950584925002393","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Context:

As Machine Learning (ML) models are increasingly used in critical decision-making software, concerns have been raised about these systems perpetuating or exacerbating existing historical biases. Consequently, there has been a growing research interest in developing methods to test for fairness and repair biases in ML software, particularly for legally protected attributes like gender, age, race.

Objectives:

In this work, we set out to repair bias for both single and multiple protected attributes (a.k.a. intersectional fairness) of pre-trained machine learning models.

Methods:

We propose a novel model- and task-agnostic debiasing method, Fair Subgroup Translation (FairST), based on fair representation learning via auto-encoders. To the best of our knowledge, this is the first approach based on the principle of Fair Representation Learning devised for post-processing bias repair.

Results:

We empirically evaluate the effectiveness of using FairST to repair a pre-trained Neural Network model used for seven classification tasks involving both single and multiple protected attributes, and benchmark its performance with state-of-the-art fairness repair methods (i.e., Learning Fair Representations, Reweighing, FairBalance and FairMask). We also investigate if the effectiveness of FairST varies when using it to repair bias of other popular ML models (namely Logistic Regression, Support Vector Machine, Gaussian Naive Bayes, Decision Tree and Random Forest).

Conclusion:

The results demonstrate that FairST consistently achieves superior single and intersectional fairness with respect to all benchmarking methods for all classification tasks considered in our empirical study. This supports the potential of using FairST for ML bias repair, and opens up a rich agenda of future work including its application to repair bias arising in tasks of a different nature such as multi-class or image-based problems.
基于潜在敏感属性翻译的机器学习偏差修复新方法
背景:随着机器学习(ML)模型越来越多地用于关键的决策软件,人们开始担心这些系统会延续或加剧现有的历史偏见。因此,人们对开发方法来测试机器学习软件中的公平性和修复偏见越来越感兴趣,特别是对性别、年龄、种族等受法律保护的属性。目的:在这项工作中,我们着手修复预训练机器学习模型的单个和多个受保护属性(也称为交叉公平性)的偏见。方法:我们提出了一种新的模型和任务不可知的去偏方法——公平子群翻译(FairST),该方法基于基于自编码器的公平表征学习。据我们所知,这是第一个基于公平代表学习原则设计的用于后处理偏见修复的方法。结果:我们通过实证评估了使用FairST修复预训练神经网络模型的有效性,该模型用于涉及单个和多个受保护属性的七个分类任务,并使用最先进的公平性修复方法(即学习公平表征、重称重、FairBalance和FairMask)对其性能进行了基准测试。我们还研究了当使用FairST来修复其他流行的ML模型(即逻辑回归、支持向量机、高斯朴素贝叶斯、决策树和随机森林)的偏差时,其有效性是否会发生变化。结论:结果表明,相对于我们实证研究中考虑的所有分类任务的所有基准测试方法,FairST始终如一地实现了优越的单一和交叉公平性。这支持了使用FairST进行机器学习偏差修复的潜力,并为未来的工作开辟了丰富的议程,包括将其应用于修复不同性质的任务(如多类或基于图像的问题)中产生的偏差。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Information and Software Technology
Information and Software Technology 工程技术-计算机:软件工程
CiteScore
9.10
自引率
7.70%
发文量
164
审稿时长
9.6 weeks
期刊介绍: Information and Software Technology is the international archival journal focusing on research and experience that contributes to the improvement of software development practices. The journal''s scope includes methods and techniques to better engineer software and manage its development. Articles submitted for review should have a clear component of software engineering or address ways to improve the engineering and management of software development. Areas covered by the journal include: • Software management, quality and metrics, • Software processes, • Software architecture, modelling, specification, design and programming • Functional and non-functional software requirements • Software testing and verification & validation • Empirical studies of all aspects of engineering and managing software development Short Communications is a new section dedicated to short papers addressing new ideas, controversial opinions, "Negative" results and much more. Read the Guide for authors for more information. The journal encourages and welcomes submissions of systematic literature studies (reviews and maps) within the scope of the journal. Information and Software Technology is the premiere outlet for systematic literature studies in software engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信