Label-aware debiased causal reasoning for Natural Language Inference

Kun Zhang , Dacao Zhang , Le Wu , Richang Hong , Ye Zhao , Meng Wang
{"title":"Label-aware debiased causal reasoning for Natural Language Inference","authors":"Kun Zhang ,&nbsp;Dacao Zhang ,&nbsp;Le Wu ,&nbsp;Richang Hong ,&nbsp;Ye Zhao ,&nbsp;Meng Wang","doi":"10.1016/j.aiopen.2024.02.001","DOIUrl":null,"url":null,"abstract":"<div><p>Recently, researchers have argued that the impressive performance of Natural Language Inference (NLI) models is highly due to the <em>spurious correlations</em> existing in training data, which makes models vulnerable and poorly generalized. Some work has made preliminary debiased attempts by developing data-driven interventions or model-level debiased learning. Despite the progress, existing debiased methods either suffered from the high cost of data annotation processing, or required elaborate design to identify biased factors. By conducting detailed investigations and data analysis, we argue that label information can provide meaningful guidance to identify these spurious correlations in training data, which has not been paid enough attention. Thus, we design a novel <em>Label-aware Debiased Causal Reasoning Network</em> (<em>LDCRN</em>). Specifically, according to the data analysis, we first build a causal graph to describe causal relations and spurious correlations in NLI. Then, we employ an NLI model (e.g., RoBERTa) to calculate total causal effect of input sentences to labels. Meanwhile, we design a novel label-aware biased module to model spurious correlations and calculate their causal effect in a fine-grained manner. The debiasing process is realized by subtracting this causal effect from total causal effect. Finally, extensive experiments over two well-known NLI datasets and multiple human-annotated challenging test sets are conducted to prove the superiority of <em>LDCRN</em>. Moreover, we have developed novel challenging test sets based on MultiNLI to facilitate the community.</p></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"5 ","pages":"Pages 70-78"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666651024000081/pdfft?md5=1863010d7dc5353ee714fa3b391ab574&pid=1-s2.0-S2666651024000081-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"AI Open","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666651024000081","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Recently, researchers have argued that the impressive performance of Natural Language Inference (NLI) models is highly due to the spurious correlations existing in training data, which makes models vulnerable and poorly generalized. Some work has made preliminary debiased attempts by developing data-driven interventions or model-level debiased learning. Despite the progress, existing debiased methods either suffered from the high cost of data annotation processing, or required elaborate design to identify biased factors. By conducting detailed investigations and data analysis, we argue that label information can provide meaningful guidance to identify these spurious correlations in training data, which has not been paid enough attention. Thus, we design a novel Label-aware Debiased Causal Reasoning Network (LDCRN). Specifically, according to the data analysis, we first build a causal graph to describe causal relations and spurious correlations in NLI. Then, we employ an NLI model (e.g., RoBERTa) to calculate total causal effect of input sentences to labels. Meanwhile, we design a novel label-aware biased module to model spurious correlations and calculate their causal effect in a fine-grained manner. The debiasing process is realized by subtracting this causal effect from total causal effect. Finally, extensive experiments over two well-known NLI datasets and multiple human-annotated challenging test sets are conducted to prove the superiority of LDCRN. Moreover, we have developed novel challenging test sets based on MultiNLI to facilitate the community.

用于自然语言推理的标签感知去标签化因果推理
近来,研究人员认为,自然语言推理(NLI)模型的出色表现在很大程度上是由于训练数据中存在的虚假相关性,这使得模型易受攻击且通用性差。一些研究通过开发数据驱动的干预或模型级去误差学习,进行了初步的去误差尝试。尽管取得了进展,但现有的去偏方法要么受制于高昂的数据注释处理成本,要么需要精心设计以识别偏差因素。通过详细调查和数据分析,我们认为标签信息可以为识别训练数据中的这些虚假相关性提供有意义的指导,而这一点尚未得到足够重视。因此,我们设计了一种新颖的标签感知偏差因果推理网络(Label-aware Debiased Causal Reasoning Network,LDCRN)。具体来说,根据数据分析,我们首先建立一个因果图来描述 NLI 中的因果关系和虚假相关性。然后,我们采用一个 NLI 模型(如 RoBERTa)来计算输入句子对标签的总因果效应。同时,我们还设计了一个新颖的标签感知偏差模块,用于对虚假相关性进行建模,并以细粒度的方式计算其因果效应。通过从总因果效应中减去这种因果效应,就实现了去伪存真的过程。最后,我们在两个著名的 NLI 数据集和多个由人类标注的挑战性测试集上进行了大量实验,以证明 LDCRN 的优越性。此外,我们还在 MultiNLI 的基础上开发了新的挑战性测试集,为社区提供便利。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
45.00
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信