{"title":"Decomposition of surprisal: Unified computational model of ERP components in language processing","authors":"Jiaxuan Li, Richard Futrell","doi":"arxiv-2409.06803","DOIUrl":null,"url":null,"abstract":"The functional interpretation of language-related ERP components has been a\ncentral debate in psycholinguistics for decades. We advance an\ninformation-theoretic model of human language processing in the brain in which\nincoming linguistic input is processed at first shallowly and later with more\ndepth, with these two kinds of information processing corresponding to distinct\nelectroencephalographic signatures. Formally, we show that the information\ncontent (surprisal) of a word in context can be decomposed into two quantities:\n(A) heuristic surprise, which signals shallow processing difficulty for a word,\nand corresponds with the N400 signal; and (B) discrepancy signal, which\nreflects the discrepancy between shallow and deep interpretations, and\ncorresponds to the P600 signal. Both of these quantities can be estimated\nstraightforwardly using modern NLP models. We validate our theory by\nsuccessfully simulating ERP patterns elicited by a variety of linguistic\nmanipulations in previously-reported experimental data from six experiments,\nwith successful novel qualitative and quantitative predictions. Our theory is\ncompatible with traditional cognitive theories assuming a `good-enough'\nheuristic interpretation stage, but with a precise information-theoretic\nformulation. The model provides an information-theoretic model of ERP\ncomponents grounded on cognitive processes, and brings us closer to a\nfully-specified neuro-computational model of language processing.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"33 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.06803","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The functional interpretation of language-related ERP components has been a
central debate in psycholinguistics for decades. We advance an
information-theoretic model of human language processing in the brain in which
incoming linguistic input is processed at first shallowly and later with more
depth, with these two kinds of information processing corresponding to distinct
electroencephalographic signatures. Formally, we show that the information
content (surprisal) of a word in context can be decomposed into two quantities:
(A) heuristic surprise, which signals shallow processing difficulty for a word,
and corresponds with the N400 signal; and (B) discrepancy signal, which
reflects the discrepancy between shallow and deep interpretations, and
corresponds to the P600 signal. Both of these quantities can be estimated
straightforwardly using modern NLP models. We validate our theory by
successfully simulating ERP patterns elicited by a variety of linguistic
manipulations in previously-reported experimental data from six experiments,
with successful novel qualitative and quantitative predictions. Our theory is
compatible with traditional cognitive theories assuming a `good-enough'
heuristic interpretation stage, but with a precise information-theoretic
formulation. The model provides an information-theoretic model of ERP
components grounded on cognitive processes, and brings us closer to a
fully-specified neuro-computational model of language processing.