深度神经网络未能捕捉人类语言的认知核心

R. Berwick
{"title":"深度神经网络未能捕捉人类语言的认知核心","authors":"R. Berwick","doi":"10.1109/iccicc53683.2021.9811297","DOIUrl":null,"url":null,"abstract":"Current deep neural networks have made remarkable advances in their ability to analyze and use natural language, with great apparent engineering success. But how well do these systems mirror the cognitive constraints associated with human language? In this talk we show that there are three essential core computations that characterize human language as an engine of human thought. One is \"digital infinity\"– the fact that we can produce an open-ended countably infinite number of sentences. The second is that sentences are hierarchically structured, rather than being arranged in a linear array. The third property is that human language computations always admit the possibility of \"displacement\" – a word or phrase can be pronounced at a place distinct from its usual location of semantic interpretation. All three properties can be shown to follow from a single, simple, recursive combinatorial operation. We provide empirical evidence for all three properties, both from concrete developmental examples as well as psycholinguistic and brain imaging experiments.What about current \"deep neural network\" systems? Although they perform very well after large-scale training, their success appears to be grounded on accurate table-lookup–memorization–without truly mirroring the three key computational principles of human language cognition. By \"stress testing\" currently available deep neural network processors, we show that they are, perhaps surprisingly very fragile when presented even with simple examples that deviate modestly from the examples on which they were trained. In particular, they fail to properly represent hierarchical structure and they cannot reliably reconstruct examples of sentences with \"displacement\" if the examples go just a bit beyond the complexity of their training set data. For example, while a deep neural network system might work on \"Which cookie did Bob want,\" it fails on, \"Which cookie did Bob want to eat.\" Such failures indicate that the neural net systems have not generalized in the same sense that children do, since children can easily handle such examples after receiving much more limited training data.","PeriodicalId":101653,"journal":{"name":"2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"The Failure of Deep Neural Networks to Capture Human Language’s Cognitive Core\",\"authors\":\"R. Berwick\",\"doi\":\"10.1109/iccicc53683.2021.9811297\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Current deep neural networks have made remarkable advances in their ability to analyze and use natural language, with great apparent engineering success. But how well do these systems mirror the cognitive constraints associated with human language? In this talk we show that there are three essential core computations that characterize human language as an engine of human thought. One is \\\"digital infinity\\\"– the fact that we can produce an open-ended countably infinite number of sentences. The second is that sentences are hierarchically structured, rather than being arranged in a linear array. The third property is that human language computations always admit the possibility of \\\"displacement\\\" – a word or phrase can be pronounced at a place distinct from its usual location of semantic interpretation. All three properties can be shown to follow from a single, simple, recursive combinatorial operation. We provide empirical evidence for all three properties, both from concrete developmental examples as well as psycholinguistic and brain imaging experiments.What about current \\\"deep neural network\\\" systems? Although they perform very well after large-scale training, their success appears to be grounded on accurate table-lookup–memorization–without truly mirroring the three key computational principles of human language cognition. By \\\"stress testing\\\" currently available deep neural network processors, we show that they are, perhaps surprisingly very fragile when presented even with simple examples that deviate modestly from the examples on which they were trained. In particular, they fail to properly represent hierarchical structure and they cannot reliably reconstruct examples of sentences with \\\"displacement\\\" if the examples go just a bit beyond the complexity of their training set data. For example, while a deep neural network system might work on \\\"Which cookie did Bob want,\\\" it fails on, \\\"Which cookie did Bob want to eat.\\\" Such failures indicate that the neural net systems have not generalized in the same sense that children do, since children can easily handle such examples after receiving much more limited training data.\",\"PeriodicalId\":101653,\"journal\":{\"name\":\"2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/iccicc53683.2021.9811297\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/iccicc53683.2021.9811297","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

当前的深度神经网络在分析和使用自然语言的能力方面取得了显著的进步,在工程上取得了巨大的成功。但是这些系统在多大程度上反映了与人类语言相关的认知限制呢?在这个演讲中,我们展示了人类语言作为人类思维引擎的三个基本核心计算。一个是“数字无限”——我们可以创造出无限数量的句子。其次,句子是分层结构的,而不是线性排列的。第三个特性是,人类语言计算总是承认“位移”的可能性——一个单词或短语的发音可能与语义解释的通常位置不同。所有这三个属性都可以通过一个简单的递归组合操作显示出来。我们从具体的发展例子以及心理语言学和脑成像实验中为这三个特性提供了经验证据。那么现在的“深度神经网络”系统呢?尽管它们在大规模训练后表现得非常好,但它们的成功似乎是建立在精确的表查找和记忆基础上的,而没有真正反映出人类语言认知的三个关键计算原则。通过“压力测试”目前可用的深度神经网络处理器,我们表明,即使是在简单的例子中,它们也可能出乎意料地非常脆弱,即使这些例子稍微偏离了它们所训练的例子。特别是,它们不能正确地表示层次结构,并且如果示例的复杂性稍微超出训练集数据的复杂性,它们就不能可靠地重建具有“位移”的句子示例。例如,虽然深度神经网络系统可能在“鲍勃想要哪块饼干”上工作,但它在“鲍勃想要吃哪块饼干”上失败了。这样的失败表明,神经网络系统没有像儿童那样泛化,因为儿童在接受更有限的训练数据后,可以很容易地处理这样的例子。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
The Failure of Deep Neural Networks to Capture Human Language’s Cognitive Core
Current deep neural networks have made remarkable advances in their ability to analyze and use natural language, with great apparent engineering success. But how well do these systems mirror the cognitive constraints associated with human language? In this talk we show that there are three essential core computations that characterize human language as an engine of human thought. One is "digital infinity"– the fact that we can produce an open-ended countably infinite number of sentences. The second is that sentences are hierarchically structured, rather than being arranged in a linear array. The third property is that human language computations always admit the possibility of "displacement" – a word or phrase can be pronounced at a place distinct from its usual location of semantic interpretation. All three properties can be shown to follow from a single, simple, recursive combinatorial operation. We provide empirical evidence for all three properties, both from concrete developmental examples as well as psycholinguistic and brain imaging experiments.What about current "deep neural network" systems? Although they perform very well after large-scale training, their success appears to be grounded on accurate table-lookup–memorization–without truly mirroring the three key computational principles of human language cognition. By "stress testing" currently available deep neural network processors, we show that they are, perhaps surprisingly very fragile when presented even with simple examples that deviate modestly from the examples on which they were trained. In particular, they fail to properly represent hierarchical structure and they cannot reliably reconstruct examples of sentences with "displacement" if the examples go just a bit beyond the complexity of their training set data. For example, while a deep neural network system might work on "Which cookie did Bob want," it fails on, "Which cookie did Bob want to eat." Such failures indicate that the neural net systems have not generalized in the same sense that children do, since children can easily handle such examples after receiving much more limited training data.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信