Understanding latent affective bias in large pre-trained neural language models

Anoop Kadan , Deepak P. , Sahely Bhadra , Manjary P. Gangan , Lajish V.L.
{"title":"Understanding latent affective bias in large pre-trained neural language models","authors":"Anoop Kadan ,&nbsp;Deepak P. ,&nbsp;Sahely Bhadra ,&nbsp;Manjary P. Gangan ,&nbsp;Lajish V.L.","doi":"10.1016/j.nlp.2024.100062","DOIUrl":null,"url":null,"abstract":"<div><p>Groundbreaking inventions and highly significant performance improvements in deep learning based Natural Language Processing are witnessed through the development of transformer based large Pre-trained Language Models (PLMs). The wide availability of unlabeled data within human generated data deluge along with self-supervised learning strategy helps to accelerate the success of large PLMs in language generation, language understanding, etc. But at the same time, latent historical bias/unfairness in human minds towards a particular gender, race, etc., encoded unintentionally/intentionally into the corpora harms and questions the utility and efficacy of large PLMs in many real-world applications, particularly for the protected groups. In this paper, we present an extensive investigation towards understanding the existence of <em>“Affective Bias”</em> in large PLMs to unveil any biased association of emotions such as <em>anger</em>, <em>fear</em>, <em>joy</em>, etc., towards a particular gender, race or religion with respect to the downstream task of textual emotion detection. We conduct our exploration of affective bias from the very initial stage of corpus level affective bias analysis by searching for imbalanced distribution of affective words within a domain, in large scale corpora that are used to pre-train and fine-tune PLMs. Later, to quantify affective bias in model predictions, we perform an extensive set of class-based and intensity-based evaluations using various bias evaluation corpora. Our results show the existence of statistically significant affective bias in the PLM based emotion detection systems, indicating biased association of certain emotions towards a particular gender, race, and religion.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100062"},"PeriodicalIF":0.0000,"publicationDate":"2024-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000104/pdfft?md5=47ed3491ca02f42caa81ecff613ee5f3&pid=1-s2.0-S2949719124000104-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Natural Language Processing Journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949719124000104","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Groundbreaking inventions and highly significant performance improvements in deep learning based Natural Language Processing are witnessed through the development of transformer based large Pre-trained Language Models (PLMs). The wide availability of unlabeled data within human generated data deluge along with self-supervised learning strategy helps to accelerate the success of large PLMs in language generation, language understanding, etc. But at the same time, latent historical bias/unfairness in human minds towards a particular gender, race, etc., encoded unintentionally/intentionally into the corpora harms and questions the utility and efficacy of large PLMs in many real-world applications, particularly for the protected groups. In this paper, we present an extensive investigation towards understanding the existence of “Affective Bias” in large PLMs to unveil any biased association of emotions such as anger, fear, joy, etc., towards a particular gender, race or religion with respect to the downstream task of textual emotion detection. We conduct our exploration of affective bias from the very initial stage of corpus level affective bias analysis by searching for imbalanced distribution of affective words within a domain, in large scale corpora that are used to pre-train and fine-tune PLMs. Later, to quantify affective bias in model predictions, we perform an extensive set of class-based and intensity-based evaluations using various bias evaluation corpora. Our results show the existence of statistically significant affective bias in the PLM based emotion detection systems, indicating biased association of certain emotions towards a particular gender, race, and religion.

Abstract Image

了解大型预训练神经语言模型中的潜在情感偏差
基于变压器的大型预训练语言模型(PLM)的开发,见证了基于深度学习的自然语言处理领域的突破性发明和非常显著的性能改进。人类产生的数据洪流中广泛存在的无标记数据以及自监督学习策略有助于加速大型 PLM 在语言生成、语言理解等方面的成功。但与此同时,人类头脑中对特定性别、种族等潜在的历史偏见/不公平,无意/有意地编码到语料库中,损害并质疑了大型 PLM 在许多现实世界应用中的实用性和有效性,尤其是对受保护群体而言。在本文中,我们对大型 PLM 中是否存在 "情感偏差 "进行了广泛的调查,以揭示在文本情感检测的下游任务中,愤怒、恐惧、喜悦等情感与特定性别、种族或宗教的关联是否存在偏差。我们从语料库情感偏差分析的初始阶段就开始探索情感偏差,在用于预训练和微调 PLM 的大型语料库中搜索某一领域中情感词的不平衡分布。随后,为了量化模型预测中的情感偏差,我们使用各种偏差评估语料库进行了大量基于类和基于强度的评估。我们的结果表明,基于 PLM 的情感检测系统在统计学上存在显著的情感偏差,表明某些情感与特定性别、种族和宗教的关联存在偏差。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信