{"title":"BiLSTM-based with Word-weight Attention for Chinese Named Entity Recognition","authors":"Ziqi Chen, Rongzhi Qi, Shui-Yan Li","doi":"10.1109/ICSESS54813.2022.9930184","DOIUrl":null,"url":null,"abstract":"Natural language processing is a hot research area in recent years. Named entity recognition is a fundamental task in natural language processing. However, Chinese named entity recognition usually suffers from weak relationship between related words and sentences resulting in recognition errors. In order to clarify the weight of different words in sentences and strengthen the dependence between character and words, we propose a named entity recognition model LSTM-WWAT based on bidirectional long-term memory network (BiLSTM) and word-weight attention(WWAT). Firstly, we add word semantic information into the character vector of the embedding layer by matching the dictionary. Secondly, we use the BiLSTM to extract the context dependent features of characters and related words. Then, the model import the hidden vector into WWAT and depend on sentences features to strengthen the word weight, so that the output will be closer to the entity annotation we want. Finally, Random Conditional Field (CRF) is used to decode the optimal coding sequence as the result of named entity recognition. Experimental results show that, compared with baseline models, our model achieves significant improvements.","PeriodicalId":265412,"journal":{"name":"2022 IEEE 13th International Conference on Software Engineering and Service Science (ICSESS)","volume":"268 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 13th International Conference on Software Engineering and Service Science (ICSESS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSESS54813.2022.9930184","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Natural language processing is a hot research area in recent years. Named entity recognition is a fundamental task in natural language processing. However, Chinese named entity recognition usually suffers from weak relationship between related words and sentences resulting in recognition errors. In order to clarify the weight of different words in sentences and strengthen the dependence between character and words, we propose a named entity recognition model LSTM-WWAT based on bidirectional long-term memory network (BiLSTM) and word-weight attention(WWAT). Firstly, we add word semantic information into the character vector of the embedding layer by matching the dictionary. Secondly, we use the BiLSTM to extract the context dependent features of characters and related words. Then, the model import the hidden vector into WWAT and depend on sentences features to strengthen the word weight, so that the output will be closer to the entity annotation we want. Finally, Random Conditional Field (CRF) is used to decode the optimal coding sequence as the result of named entity recognition. Experimental results show that, compared with baseline models, our model achieves significant improvements.