{"title":"A Deep Attention Network for Chinese Word Segment","authors":"Lanxin Li, Ping Gong, L. Ji","doi":"10.1145/3318299.3318351","DOIUrl":null,"url":null,"abstract":"Character-level sequence label tagging is the most efficient way to solve unknown words problem for Chinese word segment. But the most widely used model, Conditional Random Fields (CRF), needs a large amount of manual design features. So it is appropriate to combine CRF and neural networks such as recurrent neural network (RNN), which is adopted in many natural language processing (NLP) tasks. However, RNN is rather slow because of the timing dependence between computations and not good at capturing local information of the sentence. In order to solve this problem, we introduce a self-attention mechanism, which completes the calculation between the different positions of the sentence with the same distance, into CWS. And we propose a deep neural network, which combines convolution neural networks and self-attention mechanism. Then, we evaluate the model on the PKU dataset and the MSR dataset. The results show that our model perform much better.","PeriodicalId":164987,"journal":{"name":"International Conference on Machine Learning and Computing","volume":"38 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Machine Learning and Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3318299.3318351","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Character-level sequence label tagging is the most efficient way to solve unknown words problem for Chinese word segment. But the most widely used model, Conditional Random Fields (CRF), needs a large amount of manual design features. So it is appropriate to combine CRF and neural networks such as recurrent neural network (RNN), which is adopted in many natural language processing (NLP) tasks. However, RNN is rather slow because of the timing dependence between computations and not good at capturing local information of the sentence. In order to solve this problem, we introduce a self-attention mechanism, which completes the calculation between the different positions of the sentence with the same distance, into CWS. And we propose a deep neural network, which combines convolution neural networks and self-attention mechanism. Then, we evaluate the model on the PKU dataset and the MSR dataset. The results show that our model perform much better.