Do ever larger octopi still amplify reporting biases? Evidence from judgments of typical colour

Q3 Environmental Science
Fangyu Liu, Julian Martin Eisenschlos, Jeremy R. Cole, Nigel Collier
{"title":"Do ever larger octopi still amplify reporting biases? Evidence from judgments of typical colour","authors":"Fangyu Liu, Julian Martin Eisenschlos, Jeremy R. Cole, Nigel Collier","doi":"10.48550/arXiv.2209.12786","DOIUrl":null,"url":null,"abstract":"Language models (LMs) trained on raw texts have no direct access to the physical world. Gordon and Van Durme (2013) point out that LMs can thus suffer from reporting bias: texts rarely report on common facts, instead focusing on the unusual aspects of a situation. If LMs are only trained on text corpora and naively memorise local co-occurrence statistics, they thus naturally would learn a biased view of the physical world. While prior studies have repeatedly verified that LMs of smaller scales (e.g., RoBERTa, GPT-2) amplify reporting bias, it remains unknown whether such trends continue when models are scaled up. We investigate reporting bias from the perspective of colour in larger language models (LLMs) such as PaLM and GPT-3. Specifically, we query LLMs for the typical colour of objects, which is one simple type of perceptually grounded physical common sense. Surprisingly, we find that LLMs significantly outperform smaller LMs in determining an object’s typical colour and more closely track human judgments, instead of overfitting to surface patterns stored in texts. This suggests that very large models of language alone are able to overcome certain types of reporting bias that are characterized by local co-occurrences.","PeriodicalId":39298,"journal":{"name":"AACL Bioflux","volume":"18 1","pages":"210-220"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"AACL Bioflux","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2209.12786","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Environmental Science","Score":null,"Total":0}
引用次数: 2

Abstract

Language models (LMs) trained on raw texts have no direct access to the physical world. Gordon and Van Durme (2013) point out that LMs can thus suffer from reporting bias: texts rarely report on common facts, instead focusing on the unusual aspects of a situation. If LMs are only trained on text corpora and naively memorise local co-occurrence statistics, they thus naturally would learn a biased view of the physical world. While prior studies have repeatedly verified that LMs of smaller scales (e.g., RoBERTa, GPT-2) amplify reporting bias, it remains unknown whether such trends continue when models are scaled up. We investigate reporting bias from the perspective of colour in larger language models (LLMs) such as PaLM and GPT-3. Specifically, we query LLMs for the typical colour of objects, which is one simple type of perceptually grounded physical common sense. Surprisingly, we find that LLMs significantly outperform smaller LMs in determining an object’s typical colour and more closely track human judgments, instead of overfitting to surface patterns stored in texts. This suggests that very large models of language alone are able to overcome certain types of reporting bias that are characterized by local co-occurrences.
更大的章鱼是否仍然会放大报告的偏见?来自典型颜色判断的证据
在原始文本上训练的语言模型(LMs)无法直接访问物理世界。Gordon和Van Durme(2013)指出,LMs因此可能遭受报道偏见:文本很少报道常见事实,而是关注情况的不寻常方面。如果LMs只在文本语料库上进行训练,并天真地记忆局部共现统计数据,那么它们自然会对物理世界产生偏见。虽然先前的研究已经反复证实,较小规模的LMs(例如RoBERTa, GPT-2)会放大报告偏差,但当模型扩大时,这种趋势是否会继续,仍不得而知。我们从更大的语言模型(llm)如PaLM和GPT-3的颜色角度研究报告偏差。具体来说,我们向llm查询对象的典型颜色,这是一种简单的基于感知的物理常识。令人惊讶的是,我们发现llm在确定物体的典型颜色方面明显优于较小的lm,并且更密切地跟踪人类的判断,而不是过度拟合存储在文本中的表面模式。这表明,非常大的语言模型本身就能够克服某些类型的报告偏差,这些偏差以局部共现为特征。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
AACL Bioflux
AACL Bioflux Environmental Science-Management, Monitoring, Policy and Law
CiteScore
1.40
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信