{"title":"作为句法限制处理认知模型的神经网络。","authors":"Suhas Arehalli, Tal Linzen","doi":"10.1162/opmi_a_00137","DOIUrl":null,"url":null,"abstract":"<p><p>Languages are governed by <i>syntactic constraints</i>-structural rules that determine which sentences are grammatical in the language. In English, one such constraint is <i>subject-verb agreement</i>, which dictates that the number of a verb must match the number of its corresponding subject: \"the dog<i>s</i> run\", but \"the dog run<i>s</i>\". While this constraint appears to be simple, in practice speakers make agreement errors, particularly when a noun phrase near the verb differs in number from the subject (for example, a speaker might produce the ungrammatical sentence \"the key to the cabinets are rusty\"). This phenomenon, referred to as <i>agreement attraction</i>, is sensitive to a wide range of properties of the sentence; no single existing model is able to generate predictions for the wide variety of materials studied in the human experimental literature. We explore the viability of neural network language models-broad-coverage systems trained to predict the next word in a corpus-as a framework for addressing this limitation. We analyze the agreement errors made by Long Short-Term Memory (LSTM) networks and compare them to those of humans. The models successfully simulate certain results, such as the so-called number asymmetry and the difference between attraction strength in grammatical and ungrammatical sentences, but failed to simulate others, such as the effect of syntactic distance or notional (conceptual) number. We further evaluate networks trained with explicit syntactic supervision, and find that this form of supervision does not always lead to more human-like syntactic behavior. Finally, we show that the corpus used to train a network significantly affects the pattern of agreement errors produced by the network, and discuss the strengths and limitations of neural networks as a tool for understanding human syntactic processing.</p>","PeriodicalId":32558,"journal":{"name":"Open Mind","volume":"8 ","pages":"558-614"},"PeriodicalIF":0.0000,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11093404/pdf/","citationCount":"0","resultStr":"{\"title\":\"Neural Networks as Cognitive Models of the Processing of Syntactic Constraints.\",\"authors\":\"Suhas Arehalli, Tal Linzen\",\"doi\":\"10.1162/opmi_a_00137\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Languages are governed by <i>syntactic constraints</i>-structural rules that determine which sentences are grammatical in the language. In English, one such constraint is <i>subject-verb agreement</i>, which dictates that the number of a verb must match the number of its corresponding subject: \\\"the dog<i>s</i> run\\\", but \\\"the dog run<i>s</i>\\\". While this constraint appears to be simple, in practice speakers make agreement errors, particularly when a noun phrase near the verb differs in number from the subject (for example, a speaker might produce the ungrammatical sentence \\\"the key to the cabinets are rusty\\\"). This phenomenon, referred to as <i>agreement attraction</i>, is sensitive to a wide range of properties of the sentence; no single existing model is able to generate predictions for the wide variety of materials studied in the human experimental literature. We explore the viability of neural network language models-broad-coverage systems trained to predict the next word in a corpus-as a framework for addressing this limitation. We analyze the agreement errors made by Long Short-Term Memory (LSTM) networks and compare them to those of humans. The models successfully simulate certain results, such as the so-called number asymmetry and the difference between attraction strength in grammatical and ungrammatical sentences, but failed to simulate others, such as the effect of syntactic distance or notional (conceptual) number. We further evaluate networks trained with explicit syntactic supervision, and find that this form of supervision does not always lead to more human-like syntactic behavior. Finally, we show that the corpus used to train a network significantly affects the pattern of agreement errors produced by the network, and discuss the strengths and limitations of neural networks as a tool for understanding human syntactic processing.</p>\",\"PeriodicalId\":32558,\"journal\":{\"name\":\"Open Mind\",\"volume\":\"8 \",\"pages\":\"558-614\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11093404/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Open Mind\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1162/opmi_a_00137\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2024/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q1\",\"JCRName\":\"Social Sciences\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Open Mind","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1162/opmi_a_00137","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/1/1 0:00:00","PubModel":"eCollection","JCR":"Q1","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 0
摘要
语言受句法制约--决定语言中哪些句子符合语法的结构规则的制约。在英语中,主谓一致就是这样一种制约,它规定动词的数目必须与相应主语的数目一致:如 "the dogs run"(狗在跑),而 "the dog runs"(狗在跑)。虽然这一约束看似简单,但实际上说话者会犯一致错误,尤其是当动词附近的名词短语与主语的数字不同时(例如,说话者可能会造出 "the key to the cabinets are rusty"(橱柜的钥匙生锈了)这一不合语法的句子)。这种现象被称为 "一致吸引"(agreement attraction),它对句子的各种属性都很敏感;现有的任何一个模型都无法对人类实验文献中研究的各种材料进行预测。我们探讨了神经网络语言模型的可行性--该模型是经过训练的广覆盖系统,可预测语料库中的下一个单词,是解决这一局限性的框架。我们分析了长短期记忆(LSTM)网络所犯的一致错误,并将其与人类的错误进行了比较。这些模型成功地模拟了某些结果,如所谓的数字不对称以及语法句和非语法句中的吸引强度差异,但却无法模拟其他结果,如句法距离或概念(概念)数字的影响。我们进一步评估了使用显式句法监督训练的网络,发现这种形式的监督并不总能带来更像人类的句法行为。最后,我们证明了用于训练网络的语料库会显著影响网络产生的一致错误模式,并讨论了神经网络作为理解人类句法处理工具的优势和局限性。
Neural Networks as Cognitive Models of the Processing of Syntactic Constraints.
Languages are governed by syntactic constraints-structural rules that determine which sentences are grammatical in the language. In English, one such constraint is subject-verb agreement, which dictates that the number of a verb must match the number of its corresponding subject: "the dogs run", but "the dog runs". While this constraint appears to be simple, in practice speakers make agreement errors, particularly when a noun phrase near the verb differs in number from the subject (for example, a speaker might produce the ungrammatical sentence "the key to the cabinets are rusty"). This phenomenon, referred to as agreement attraction, is sensitive to a wide range of properties of the sentence; no single existing model is able to generate predictions for the wide variety of materials studied in the human experimental literature. We explore the viability of neural network language models-broad-coverage systems trained to predict the next word in a corpus-as a framework for addressing this limitation. We analyze the agreement errors made by Long Short-Term Memory (LSTM) networks and compare them to those of humans. The models successfully simulate certain results, such as the so-called number asymmetry and the difference between attraction strength in grammatical and ungrammatical sentences, but failed to simulate others, such as the effect of syntactic distance or notional (conceptual) number. We further evaluate networks trained with explicit syntactic supervision, and find that this form of supervision does not always lead to more human-like syntactic behavior. Finally, we show that the corpus used to train a network significantly affects the pattern of agreement errors produced by the network, and discuss the strengths and limitations of neural networks as a tool for understanding human syntactic processing.