{"title":"大型语言模型中基于上下文与基于结构的相对从句消歧","authors":"Elsayed Issa , Noureddine Atouf","doi":"10.1016/j.procs.2024.10.217","DOIUrl":null,"url":null,"abstract":"<div><div>This work investigates the processing behavior of large language models (LLMs) in sentences involving ambiguous relative clauses (RCs). We are particularly interested in unravelling attachment preferences of LLMs in disambiguating RCs (complementizer phrases CPs modifying a genitival phrase), which are either semantically (context-biased) or syntactically (structure-biased) associated to one of the preceding NP referents. A low interpretation of the RC occurs when it is joined to the local NP (low attachment). A high interpretation is provided when the RC modifies the distant NP (high attachment). We create a small dataset of parallel low- and high-attachment sentences. We use zero-shot prompting to evaluate a set of LLMs based on insights from psycholinguistic experiments. Our results show variability in the performance of some models that favor low attachment (semantically-related meanings in the CP) while other models can resolve ambiguity by choosing high-attachment (structure-biased CPs). The findings are discussed in light of directing future experimental studies to consider a comparative paradigm encompassing both multi-modal LLMs and human subjects.</div></div>","PeriodicalId":20465,"journal":{"name":"Procedia Computer Science","volume":"244 ","pages":"Pages 425-431"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Context-biased vs. structure-biased disambiguation of relative clauses in large language models\",\"authors\":\"Elsayed Issa , Noureddine Atouf\",\"doi\":\"10.1016/j.procs.2024.10.217\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This work investigates the processing behavior of large language models (LLMs) in sentences involving ambiguous relative clauses (RCs). We are particularly interested in unravelling attachment preferences of LLMs in disambiguating RCs (complementizer phrases CPs modifying a genitival phrase), which are either semantically (context-biased) or syntactically (structure-biased) associated to one of the preceding NP referents. A low interpretation of the RC occurs when it is joined to the local NP (low attachment). A high interpretation is provided when the RC modifies the distant NP (high attachment). We create a small dataset of parallel low- and high-attachment sentences. We use zero-shot prompting to evaluate a set of LLMs based on insights from psycholinguistic experiments. Our results show variability in the performance of some models that favor low attachment (semantically-related meanings in the CP) while other models can resolve ambiguity by choosing high-attachment (structure-biased CPs). The findings are discussed in light of directing future experimental studies to consider a comparative paradigm encompassing both multi-modal LLMs and human subjects.</div></div>\",\"PeriodicalId\":20465,\"journal\":{\"name\":\"Procedia Computer Science\",\"volume\":\"244 \",\"pages\":\"Pages 425-431\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Procedia Computer Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1877050924030187\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Procedia Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1877050924030187","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Context-biased vs. structure-biased disambiguation of relative clauses in large language models
This work investigates the processing behavior of large language models (LLMs) in sentences involving ambiguous relative clauses (RCs). We are particularly interested in unravelling attachment preferences of LLMs in disambiguating RCs (complementizer phrases CPs modifying a genitival phrase), which are either semantically (context-biased) or syntactically (structure-biased) associated to one of the preceding NP referents. A low interpretation of the RC occurs when it is joined to the local NP (low attachment). A high interpretation is provided when the RC modifies the distant NP (high attachment). We create a small dataset of parallel low- and high-attachment sentences. We use zero-shot prompting to evaluate a set of LLMs based on insights from psycholinguistic experiments. Our results show variability in the performance of some models that favor low attachment (semantically-related meanings in the CP) while other models can resolve ambiguity by choosing high-attachment (structure-biased CPs). The findings are discussed in light of directing future experimental studies to consider a comparative paradigm encompassing both multi-modal LLMs and human subjects.