{"title":"A Model of Spoken Language Understanding Combining with Multi-Head Self-Attention","authors":"Dafei Lin, Jiangfeng Zhou, Xinlai Xing, Xiaochuan Zhang","doi":"10.1109/ACAIT56212.2022.10137905","DOIUrl":null,"url":null,"abstract":"Spoken Language Understanding (SLU) is a very important module in intelligent dialogue systems. It is usually constructed based on a bi-directional long and short-term memory network (BiLSTM). It has some shortcomings, such as relative single representation of feature space and fuzzy semantic features. For this reason, this study constructs a SLU model which combines the temporal characteristics of context and the characteristics of multi-layer representation space. The model combines a bi-directional long and short-term memory network and a multi-head self-attention to extract different feature information of contextual temporal features and multisemantic representation space of the text, respectively; then, the two features are fused using a residual linking method to enhance the features of word dependence at different locations within the text; meanwhile, the gate mechanism is then used to enable the intent detection task to establish an influence relationship on the slot filling task. Finally, the SNIPS dataset, the ATIS dataset, and the slot-gated model are selected for comparison experiments. The slot filling F1 value is increased by 4.14% and 1.1% respectively, and the accuracy of semantic framework is increased by 4.25% and 2.50% respectively. The results show the effectiveness of the model of SLU task.","PeriodicalId":398228,"journal":{"name":"2022 6th Asian Conference on Artificial Intelligence Technology (ACAIT)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 6th Asian Conference on Artificial Intelligence Technology (ACAIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACAIT56212.2022.10137905","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Spoken Language Understanding (SLU) is a very important module in intelligent dialogue systems. It is usually constructed based on a bi-directional long and short-term memory network (BiLSTM). It has some shortcomings, such as relative single representation of feature space and fuzzy semantic features. For this reason, this study constructs a SLU model which combines the temporal characteristics of context and the characteristics of multi-layer representation space. The model combines a bi-directional long and short-term memory network and a multi-head self-attention to extract different feature information of contextual temporal features and multisemantic representation space of the text, respectively; then, the two features are fused using a residual linking method to enhance the features of word dependence at different locations within the text; meanwhile, the gate mechanism is then used to enable the intent detection task to establish an influence relationship on the slot filling task. Finally, the SNIPS dataset, the ATIS dataset, and the slot-gated model are selected for comparison experiments. The slot filling F1 value is increased by 4.14% and 1.1% respectively, and the accuracy of semantic framework is increased by 4.25% and 2.50% respectively. The results show the effectiveness of the model of SLU task.