{"title":"Evaluating the Cognitively-Related Productivity of a Universal Dependency Parser","authors":"Sagar Indurkhya, R. Berwick","doi":"10.1109/ICCICC53683.2021.9811322","DOIUrl":null,"url":null,"abstract":"A key goal of cognitive computing is to correctly model human language. Recently, much has been made of the ability of deep neural nets trained on huge datasets to precisely parse sentences. But do these systems truly incorporate human knowledge of language? In this paper we apply a standard linguistic methodology, transformational analysis, to determine whether this claim is accurate. On this view, if a deep net parser operates properly on one kind of sentence, it should also work correctly on its transformed counterpart. Applying this to a standard set of statement-question transformed sentence pairs, we find that a state of the art neural network system does not replicate human behavior and makes numerous errors. We suggest that this kind of test is more relevant for highlighting what deep neural networks can and cannot do with respect to human language.","PeriodicalId":101653,"journal":{"name":"2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCICC53683.2021.9811322","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
A key goal of cognitive computing is to correctly model human language. Recently, much has been made of the ability of deep neural nets trained on huge datasets to precisely parse sentences. But do these systems truly incorporate human knowledge of language? In this paper we apply a standard linguistic methodology, transformational analysis, to determine whether this claim is accurate. On this view, if a deep net parser operates properly on one kind of sentence, it should also work correctly on its transformed counterpart. Applying this to a standard set of statement-question transformed sentence pairs, we find that a state of the art neural network system does not replicate human behavior and makes numerous errors. We suggest that this kind of test is more relevant for highlighting what deep neural networks can and cannot do with respect to human language.