实现公平决策:消除预训练模型缺陷的新型表示方法

IF 6.7 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Junheng He , Nankai Lin , Qifeng Bai , Haoyu Liang , Dong Zhou , Aimin Yang
{"title":"实现公平决策:消除预训练模型缺陷的新型表示方法","authors":"Junheng He ,&nbsp;Nankai Lin ,&nbsp;Qifeng Bai ,&nbsp;Haoyu Liang ,&nbsp;Dong Zhou ,&nbsp;Aimin Yang","doi":"10.1016/j.dss.2024.114208","DOIUrl":null,"url":null,"abstract":"<div><p>Pretrained language models (PLMs) are frequently employed in Decision Support Systems (DSSs) due to their strong performance. However, recent studies have revealed that these PLMs can exhibit social biases, leading to unfair decisions that harm vulnerable groups. Sensitive information contained in sentences from training data is the primary source of bias. Previously proposed debiasing methods based on contrastive disentanglement have proven highly effective. In these methods, PLMs can disentangle sensitive information from non-sensitive information in sentence embedding, and then adapts non-sensitive information only for downstream tasks. Such approaches hinge on having good sentence embedding as input. However, recent research found that most non-fine-tuned PLMs such as BERT produce poor sentence embedding. Disentangling based on these embedding will lead to unsatisfactory debiasing results. Taking a finer-grained perspective, we propose PCFR (Prompt and Contrastive-based Fair Representation), a novel disentanglement method integrating prompt and contrastive learning to debias PLMs. We employ prompt learning to represent information as sensitive embedding and subsequently apply contrastive learning to contrast these information embedding rather than the sentence embedding. PCFR encourages similarity among different non-sensitive information embedding and dissimilarity between sensitive and non-sensitive information embedding. We mitigate gender and religion biases in two prominent PLMs, namely BERT and GPT-2. To comprehensively assess debiasing efficacy of PCFR, we employ multiple fairness metrics. Experimental results consistently demonstrate the superior performance of PCFR compared to representative baseline methods. Additionally, when applied to specific downstream decision tasks, PCFR not only shows strong de-biasing capability but also significantly preserves task performance.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"181 ","pages":"Article 114208"},"PeriodicalIF":6.7000,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards fair decision: A novel representation method for debiasing pre-trained models\",\"authors\":\"Junheng He ,&nbsp;Nankai Lin ,&nbsp;Qifeng Bai ,&nbsp;Haoyu Liang ,&nbsp;Dong Zhou ,&nbsp;Aimin Yang\",\"doi\":\"10.1016/j.dss.2024.114208\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Pretrained language models (PLMs) are frequently employed in Decision Support Systems (DSSs) due to their strong performance. However, recent studies have revealed that these PLMs can exhibit social biases, leading to unfair decisions that harm vulnerable groups. Sensitive information contained in sentences from training data is the primary source of bias. Previously proposed debiasing methods based on contrastive disentanglement have proven highly effective. In these methods, PLMs can disentangle sensitive information from non-sensitive information in sentence embedding, and then adapts non-sensitive information only for downstream tasks. Such approaches hinge on having good sentence embedding as input. However, recent research found that most non-fine-tuned PLMs such as BERT produce poor sentence embedding. Disentangling based on these embedding will lead to unsatisfactory debiasing results. Taking a finer-grained perspective, we propose PCFR (Prompt and Contrastive-based Fair Representation), a novel disentanglement method integrating prompt and contrastive learning to debias PLMs. We employ prompt learning to represent information as sensitive embedding and subsequently apply contrastive learning to contrast these information embedding rather than the sentence embedding. PCFR encourages similarity among different non-sensitive information embedding and dissimilarity between sensitive and non-sensitive information embedding. We mitigate gender and religion biases in two prominent PLMs, namely BERT and GPT-2. To comprehensively assess debiasing efficacy of PCFR, we employ multiple fairness metrics. Experimental results consistently demonstrate the superior performance of PCFR compared to representative baseline methods. Additionally, when applied to specific downstream decision tasks, PCFR not only shows strong de-biasing capability but also significantly preserves task performance.</p></div>\",\"PeriodicalId\":55181,\"journal\":{\"name\":\"Decision Support Systems\",\"volume\":\"181 \",\"pages\":\"Article 114208\"},\"PeriodicalIF\":6.7000,\"publicationDate\":\"2024-03-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Decision Support Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167923624000411\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Decision Support Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167923624000411","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

预训练语言模型(PLM)因其强大的性能,经常被用于决策支持系统(DSS)中。然而,最近的研究发现,这些 PLM 可能会表现出社会偏见,从而导致不公平的决策,损害弱势群体的利益。训练数据中的句子所包含的敏感信息是偏见的主要来源。之前提出的基于对比分解的去偏差方法被证明非常有效。在这些方法中,PLM 可以将句子嵌入中的敏感信息与非敏感信息分离开来,然后只在下游任务中使用非敏感信息。这些方法的前提是要有良好的句子嵌入作为输入。然而,最近的研究发现,大多数非微调 PLM(如 BERT)产生的句子嵌入效果不佳。根据这些嵌入进行解刨会导致令人不满意的解刨结果。从更精细的角度出发,我们提出了 PCFR(基于提示和对比的公平表征),这是一种整合了提示学习和对比学习的新型解缠方法,可用于去除 PLM。我们利用提示学习将信息表述为敏感嵌入,然后应用对比学习来对比这些信息嵌入而不是句子嵌入。PCFR 鼓励不同非敏感信息嵌入之间的相似性以及敏感和非敏感信息嵌入之间的差异性。我们在两个著名的 PLM(即 BERT 和 GPT-2)中减轻了性别和宗教偏见。为了全面评估 PCFR 的消除偏差效果,我们采用了多种公平性指标。实验结果一致表明,与具有代表性的基线方法相比,PCFR 的性能更加优越。此外,当应用于特定的下游决策任务时,PCFR 不仅显示出强大的去偏差能力,还能显著保持任务性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Towards fair decision: A novel representation method for debiasing pre-trained models

Pretrained language models (PLMs) are frequently employed in Decision Support Systems (DSSs) due to their strong performance. However, recent studies have revealed that these PLMs can exhibit social biases, leading to unfair decisions that harm vulnerable groups. Sensitive information contained in sentences from training data is the primary source of bias. Previously proposed debiasing methods based on contrastive disentanglement have proven highly effective. In these methods, PLMs can disentangle sensitive information from non-sensitive information in sentence embedding, and then adapts non-sensitive information only for downstream tasks. Such approaches hinge on having good sentence embedding as input. However, recent research found that most non-fine-tuned PLMs such as BERT produce poor sentence embedding. Disentangling based on these embedding will lead to unsatisfactory debiasing results. Taking a finer-grained perspective, we propose PCFR (Prompt and Contrastive-based Fair Representation), a novel disentanglement method integrating prompt and contrastive learning to debias PLMs. We employ prompt learning to represent information as sensitive embedding and subsequently apply contrastive learning to contrast these information embedding rather than the sentence embedding. PCFR encourages similarity among different non-sensitive information embedding and dissimilarity between sensitive and non-sensitive information embedding. We mitigate gender and religion biases in two prominent PLMs, namely BERT and GPT-2. To comprehensively assess debiasing efficacy of PCFR, we employ multiple fairness metrics. Experimental results consistently demonstrate the superior performance of PCFR compared to representative baseline methods. Additionally, when applied to specific downstream decision tasks, PCFR not only shows strong de-biasing capability but also significantly preserves task performance.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Decision Support Systems
Decision Support Systems 工程技术-计算机:人工智能
CiteScore
14.70
自引率
6.70%
发文量
119
审稿时长
13 months
期刊介绍: The common thread of articles published in Decision Support Systems is their relevance to theoretical and technical issues in the support of enhanced decision making. The areas addressed may include foundations, functionality, interfaces, implementation, impacts, and evaluation of decision support systems (DSSs).
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信