{"title":"What's Surprising About Surprisal.","authors":"Sophie Slaats, Andrea E Martin","doi":"10.1007/s42113-025-00237-9","DOIUrl":null,"url":null,"abstract":"<p><p>In the computational and experimental psycholinguistic literature, the mechanisms behind syntactic structure building (e.g., combining words into phrases and sentences) are the subject of considerable debate. Much experimental work has shown that surprisal is a good predictor of human behavioral and neural data. These findings have led some authors to model language comprehension in a purely probabilistic way. In this paper, we use simulation to exemplify why surprisal works so well to model human data and to illustrate why exclusive reliance on it can be problematic for the development of mechanistic theories of language comprehension, particularly those with emphasis on meaning composition. Rather than arguing for the importance of structural or probabilistic information to the exclusion or exhaustion of the other, we argue more emphasis should be placed on understanding how the brain leverages both types of information (viz., statistical and structured). We propose that probabilistic information is an important <i>cue</i> to the structure in the message, but is not a substitute for the structure itself-neither computationally, formally, nor conceptually. Surprisal and other probabilistic metrics must play a key role as theoretical objects in any explanatory mechanistic theory of language processing, but that role remains in the service of the brain's goal of constructing structured meaning from sensory input.</p><p><strong>Supplementary information: </strong>The online version contains supplementary material available at 10.1007/s42113-025-00237-9.</p>","PeriodicalId":72660,"journal":{"name":"Computational brain & behavior","volume":"8 2","pages":"233-248"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12125142/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computational brain & behavior","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s42113-025-00237-9","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/21 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In the computational and experimental psycholinguistic literature, the mechanisms behind syntactic structure building (e.g., combining words into phrases and sentences) are the subject of considerable debate. Much experimental work has shown that surprisal is a good predictor of human behavioral and neural data. These findings have led some authors to model language comprehension in a purely probabilistic way. In this paper, we use simulation to exemplify why surprisal works so well to model human data and to illustrate why exclusive reliance on it can be problematic for the development of mechanistic theories of language comprehension, particularly those with emphasis on meaning composition. Rather than arguing for the importance of structural or probabilistic information to the exclusion or exhaustion of the other, we argue more emphasis should be placed on understanding how the brain leverages both types of information (viz., statistical and structured). We propose that probabilistic information is an important cue to the structure in the message, but is not a substitute for the structure itself-neither computationally, formally, nor conceptually. Surprisal and other probabilistic metrics must play a key role as theoretical objects in any explanatory mechanistic theory of language processing, but that role remains in the service of the brain's goal of constructing structured meaning from sensory input.
Supplementary information: The online version contains supplementary material available at 10.1007/s42113-025-00237-9.