Decomposition of surprisal: Unified computational model of ERP components in language processing

Jiaxuan Li, Richard Futrell
{"title":"Decomposition of surprisal: Unified computational model of ERP components in language processing","authors":"Jiaxuan Li, Richard Futrell","doi":"arxiv-2409.06803","DOIUrl":null,"url":null,"abstract":"The functional interpretation of language-related ERP components has been a\ncentral debate in psycholinguistics for decades. We advance an\ninformation-theoretic model of human language processing in the brain in which\nincoming linguistic input is processed at first shallowly and later with more\ndepth, with these two kinds of information processing corresponding to distinct\nelectroencephalographic signatures. Formally, we show that the information\ncontent (surprisal) of a word in context can be decomposed into two quantities:\n(A) heuristic surprise, which signals shallow processing difficulty for a word,\nand corresponds with the N400 signal; and (B) discrepancy signal, which\nreflects the discrepancy between shallow and deep interpretations, and\ncorresponds to the P600 signal. Both of these quantities can be estimated\nstraightforwardly using modern NLP models. We validate our theory by\nsuccessfully simulating ERP patterns elicited by a variety of linguistic\nmanipulations in previously-reported experimental data from six experiments,\nwith successful novel qualitative and quantitative predictions. Our theory is\ncompatible with traditional cognitive theories assuming a `good-enough'\nheuristic interpretation stage, but with a precise information-theoretic\nformulation. The model provides an information-theoretic model of ERP\ncomponents grounded on cognitive processes, and brings us closer to a\nfully-specified neuro-computational model of language processing.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"33 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.06803","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The functional interpretation of language-related ERP components has been a central debate in psycholinguistics for decades. We advance an information-theoretic model of human language processing in the brain in which incoming linguistic input is processed at first shallowly and later with more depth, with these two kinds of information processing corresponding to distinct electroencephalographic signatures. Formally, we show that the information content (surprisal) of a word in context can be decomposed into two quantities: (A) heuristic surprise, which signals shallow processing difficulty for a word, and corresponds with the N400 signal; and (B) discrepancy signal, which reflects the discrepancy between shallow and deep interpretations, and corresponds to the P600 signal. Both of these quantities can be estimated straightforwardly using modern NLP models. We validate our theory by successfully simulating ERP patterns elicited by a variety of linguistic manipulations in previously-reported experimental data from six experiments, with successful novel qualitative and quantitative predictions. Our theory is compatible with traditional cognitive theories assuming a `good-enough' heuristic interpretation stage, but with a precise information-theoretic formulation. The model provides an information-theoretic model of ERP components grounded on cognitive processes, and brings us closer to a fully-specified neuro-computational model of language processing.
惊喜的分解:语言处理中ERP成分的统一计算模型
几十年来,与语言相关的 ERP 成分的功能解释一直是心理语言学的核心争论点。我们提出了一个大脑中人类语言处理的信息理论模型,在这个模型中,输入的语言输入首先经过浅层处理,然后再经过更深层的处理,这两种信息处理对应于不同的脑电图特征。从形式上看,我们发现单词在上下文中的信息含量(惊奇)可以分解为两个量:(A)启发式惊奇,它表示单词的浅层处理难度,与 N400 信号相对应;(B)差异信号,它反映了浅层和深层解释之间的差异,与 P600 信号相对应。这两个量都可以使用现代 NLP 模型直接估算。我们通过在之前报告的六个实验数据中成功模拟了各种语言操纵所引发的ERP模式,并成功地进行了新颖的定性和定量预测,从而验证了我们的理论。我们的理论与假定 "足够好 "的启发式解释阶段的传统认知理论兼容,但具有精确的信息理论表述。该模型为基于认知过程的ERP成分提供了一个信息理论模型,并使我们更接近于语言处理的神经计算模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信