{"title":"Sentence processing by humans and machines: Large language models as a tool to better understand human reading.","authors":"Nikki G Kaye, Peter C Gordon","doi":"10.3758/s13423-025-02756-9","DOIUrl":null,"url":null,"abstract":"<p><p>Online measures of reading have been studied with the goal of understanding how humans process language incrementally as they progress through a text. A focus of this research has been on pinpointing how the context of a word influences its processing. Quantitatively measuring the effects of context has proven difficult but with advances in artificial intelligence, large language models (LLMs) are more capable of generating humanlike language, drawing solely on information about the probabilistic relationships of units of language (e.g., words) occurring together. LLMs can be used to estimate the probability of any word in the model's vocabulary occurring as the next word in a given context. These next-word probabilities can be used in the calculation of information theoretic metrics, such as entropy and surprisal, which can be assessed as measures of word-by-word processing load. This is done by analyzing whether entropy and surprisal derived from language models predict variance in online measures of human reading comprehension (e.g., eye-movement, self-paced reading, ERP data). The present review synthesizes empirical findings on this topic and evaluates their methodological and theoretical implications.</p>","PeriodicalId":20763,"journal":{"name":"Psychonomic Bulletin & Review","volume":" ","pages":""},"PeriodicalIF":3.0000,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Psychonomic Bulletin & Review","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.3758/s13423-025-02756-9","RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, EXPERIMENTAL","Score":null,"Total":0}
引用次数: 0
Abstract
Online measures of reading have been studied with the goal of understanding how humans process language incrementally as they progress through a text. A focus of this research has been on pinpointing how the context of a word influences its processing. Quantitatively measuring the effects of context has proven difficult but with advances in artificial intelligence, large language models (LLMs) are more capable of generating humanlike language, drawing solely on information about the probabilistic relationships of units of language (e.g., words) occurring together. LLMs can be used to estimate the probability of any word in the model's vocabulary occurring as the next word in a given context. These next-word probabilities can be used in the calculation of information theoretic metrics, such as entropy and surprisal, which can be assessed as measures of word-by-word processing load. This is done by analyzing whether entropy and surprisal derived from language models predict variance in online measures of human reading comprehension (e.g., eye-movement, self-paced reading, ERP data). The present review synthesizes empirical findings on this topic and evaluates their methodological and theoretical implications.
期刊介绍:
The journal provides coverage spanning a broad spectrum of topics in all areas of experimental psychology. The journal is primarily dedicated to the publication of theory and review articles and brief reports of outstanding experimental work. Areas of coverage include cognitive psychology broadly construed, including but not limited to action, perception, & attention, language, learning & memory, reasoning & decision making, and social cognition. We welcome submissions that approach these issues from a variety of perspectives such as behavioral measurements, comparative psychology, development, evolutionary psychology, genetics, neuroscience, and quantitative/computational modeling. We particularly encourage integrative research that crosses traditional content and methodological boundaries.