Jingxin Liu , Mengzhe Sun , Wenhao Zhang , Gengquan Xie , Yongxia Jing , Xiulai Li , Zhaoxin Shi
{"title":"DAE-NER: Dual-channel attention enhancement for Chinese named entity recognition","authors":"Jingxin Liu , Mengzhe Sun , Wenhao Zhang , Gengquan Xie , Yongxia Jing , Xiulai Li , Zhaoxin Shi","doi":"10.1016/j.csl.2023.101581","DOIUrl":null,"url":null,"abstract":"<div><p>Named Entity Recognition (NER) is an important component of Natural Language Processing (NLP) and is a fundamental yet challenging task in text analysis. Recently, NER models for Chinese-language characters have received considerable attention. Owing to the complexity and ambiguity of the Chinese language, the same semantic features have different levels of importance in different contexts. However, existing literature on Chinese Named Entity recognition (CNER) does not capture this difference in importance. To tackle this problem, we propose a new method, referred to as Dual-channel Attention Enhancement for Chinese Named Entity Recognition (DAE-NER). Specifically, we design compression and decompression mechanisms to adapt Chinese language characters to different contexts. By adjusting the weight of the semantic feature vector, the semantic weight is reconstructed to alleviate the interference of contextual differences in semantics. Moreover, in order to enhance the semantic representation of the different granularities in Chinese text, we design attention enhancement modules at the character and sentence levels. These modules dynamically learn the differences in semantic features to enhance important semantic representations in different dimensions. Extensive experiments on four benchmark datasets, namely MSRA, People Daily, Resume, and Weibo, have demonstrated that the proposed DAE-NER can effectively improve the overall performance of CNER.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":3.1000,"publicationDate":"2023-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Speech and Language","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0885230823001006","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Named Entity Recognition (NER) is an important component of Natural Language Processing (NLP) and is a fundamental yet challenging task in text analysis. Recently, NER models for Chinese-language characters have received considerable attention. Owing to the complexity and ambiguity of the Chinese language, the same semantic features have different levels of importance in different contexts. However, existing literature on Chinese Named Entity recognition (CNER) does not capture this difference in importance. To tackle this problem, we propose a new method, referred to as Dual-channel Attention Enhancement for Chinese Named Entity Recognition (DAE-NER). Specifically, we design compression and decompression mechanisms to adapt Chinese language characters to different contexts. By adjusting the weight of the semantic feature vector, the semantic weight is reconstructed to alleviate the interference of contextual differences in semantics. Moreover, in order to enhance the semantic representation of the different granularities in Chinese text, we design attention enhancement modules at the character and sentence levels. These modules dynamically learn the differences in semantic features to enhance important semantic representations in different dimensions. Extensive experiments on four benchmark datasets, namely MSRA, People Daily, Resume, and Weibo, have demonstrated that the proposed DAE-NER can effectively improve the overall performance of CNER.
期刊介绍:
Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language.
The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with complex models of speech and language processing has become feasible. Such research is often carried out somewhat separately by practitioners of artificial intelligence, computer science, electronic engineering, information retrieval, linguistics, phonetics, or psychology.