{"title":"基于BERT预训练模型的汉语意图识别研究","authors":"P. Zhang, Li Huang","doi":"10.1145/3395260.3395274","DOIUrl":null,"url":null,"abstract":"As a sub-task in natural language understanding, intent recognition research plays an important role in it. The accuracy of intent recognition is directly related to the performance of semantic slot filling, the choice of data set, and the research that will affect subsequent dialogue systems. Considering the diversity in text representation, traditional machine learning has been unable to accurately understand the deep meaning of user texts. This paper uses a BERT pre-trained model in deep learning based on Chinese text knots, and then adds a linear classification to it. Using the downstream classification task to fine-tune the pre-trained model so that the entire model together maximizes the performance of the downstream task. This paper performs domain intent classification experiments on the Chinese text THUCNews dataset.Compared with recurrent neural network(RNN) and convolutional neural network(CNN) methods, this method can improve performance by 3 percentage points. Experimental results show that the BERT pre-trained model can provide better accuracy and recall of Chinese news text domain intent classification.","PeriodicalId":103490,"journal":{"name":"Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence","volume":"82 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Research on Chinese Intent Recognition Based on BERT pre-trained model\",\"authors\":\"P. Zhang, Li Huang\",\"doi\":\"10.1145/3395260.3395274\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As a sub-task in natural language understanding, intent recognition research plays an important role in it. The accuracy of intent recognition is directly related to the performance of semantic slot filling, the choice of data set, and the research that will affect subsequent dialogue systems. Considering the diversity in text representation, traditional machine learning has been unable to accurately understand the deep meaning of user texts. This paper uses a BERT pre-trained model in deep learning based on Chinese text knots, and then adds a linear classification to it. Using the downstream classification task to fine-tune the pre-trained model so that the entire model together maximizes the performance of the downstream task. This paper performs domain intent classification experiments on the Chinese text THUCNews dataset.Compared with recurrent neural network(RNN) and convolutional neural network(CNN) methods, this method can improve performance by 3 percentage points. Experimental results show that the BERT pre-trained model can provide better accuracy and recall of Chinese news text domain intent classification.\",\"PeriodicalId\":103490,\"journal\":{\"name\":\"Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence\",\"volume\":\"82 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-04-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3395260.3395274\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3395260.3395274","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Research on Chinese Intent Recognition Based on BERT pre-trained model
As a sub-task in natural language understanding, intent recognition research plays an important role in it. The accuracy of intent recognition is directly related to the performance of semantic slot filling, the choice of data set, and the research that will affect subsequent dialogue systems. Considering the diversity in text representation, traditional machine learning has been unable to accurately understand the deep meaning of user texts. This paper uses a BERT pre-trained model in deep learning based on Chinese text knots, and then adds a linear classification to it. Using the downstream classification task to fine-tune the pre-trained model so that the entire model together maximizes the performance of the downstream task. This paper performs domain intent classification experiments on the Chinese text THUCNews dataset.Compared with recurrent neural network(RNN) and convolutional neural network(CNN) methods, this method can improve performance by 3 percentage points. Experimental results show that the BERT pre-trained model can provide better accuracy and recall of Chinese news text domain intent classification.