{"title":"Context-biased vs. structure-biased disambiguation of relative clauses in large language models","authors":"Elsayed Issa , Noureddine Atouf","doi":"10.1016/j.procs.2024.10.217","DOIUrl":null,"url":null,"abstract":"<div><div>This work investigates the processing behavior of large language models (LLMs) in sentences involving ambiguous relative clauses (RCs). We are particularly interested in unravelling attachment preferences of LLMs in disambiguating RCs (complementizer phrases CPs modifying a genitival phrase), which are either semantically (context-biased) or syntactically (structure-biased) associated to one of the preceding NP referents. A low interpretation of the RC occurs when it is joined to the local NP (low attachment). A high interpretation is provided when the RC modifies the distant NP (high attachment). We create a small dataset of parallel low- and high-attachment sentences. We use zero-shot prompting to evaluate a set of LLMs based on insights from psycholinguistic experiments. Our results show variability in the performance of some models that favor low attachment (semantically-related meanings in the CP) while other models can resolve ambiguity by choosing high-attachment (structure-biased CPs). The findings are discussed in light of directing future experimental studies to consider a comparative paradigm encompassing both multi-modal LLMs and human subjects.</div></div>","PeriodicalId":20465,"journal":{"name":"Procedia Computer Science","volume":"244 ","pages":"Pages 425-431"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Procedia Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1877050924030187","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This work investigates the processing behavior of large language models (LLMs) in sentences involving ambiguous relative clauses (RCs). We are particularly interested in unravelling attachment preferences of LLMs in disambiguating RCs (complementizer phrases CPs modifying a genitival phrase), which are either semantically (context-biased) or syntactically (structure-biased) associated to one of the preceding NP referents. A low interpretation of the RC occurs when it is joined to the local NP (low attachment). A high interpretation is provided when the RC modifies the distant NP (high attachment). We create a small dataset of parallel low- and high-attachment sentences. We use zero-shot prompting to evaluate a set of LLMs based on insights from psycholinguistic experiments. Our results show variability in the performance of some models that favor low attachment (semantically-related meanings in the CP) while other models can resolve ambiguity by choosing high-attachment (structure-biased CPs). The findings are discussed in light of directing future experimental studies to consider a comparative paradigm encompassing both multi-modal LLMs and human subjects.