{"title":"Domain-Specific Chinese Transformer-XL Language Model with Part-of-Speech Information","authors":"Huaichang Qu, Haifeng Zhao, Xin Wang","doi":"10.1109/CIS52066.2020.00026","DOIUrl":null,"url":null,"abstract":"Language models hope to use more context to predict the next word. However, not all words in the context are related to the next word and are effective for prediction. The language model based on the attention mechanism can select more useful word representations from the context and efficiently use long-term historical information. In this paper, we will apply Transformer-XL language model to Chinese automatic speech recognition in a specific domain. We add part-of-speech information for domain adaptation. First, we construct a Chinese corpus dataset in a specific domain. And by collecting common vocabulary and extracting new words in the domain, we also construct a domain vocabulary. Then, the Chinese word boundary information is added to the Transformer-XL language model to make the model can better adapt to the characteristics of the domain. Finally, our experimental results show that the method is effective on the dataset we provided. It can further reduce the Character Error Rate (CER) in speech recognition.","PeriodicalId":106959,"journal":{"name":"2020 16th International Conference on Computational Intelligence and Security (CIS)","volume":"698 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 16th International Conference on Computational Intelligence and Security (CIS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIS52066.2020.00026","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Language models hope to use more context to predict the next word. However, not all words in the context are related to the next word and are effective for prediction. The language model based on the attention mechanism can select more useful word representations from the context and efficiently use long-term historical information. In this paper, we will apply Transformer-XL language model to Chinese automatic speech recognition in a specific domain. We add part-of-speech information for domain adaptation. First, we construct a Chinese corpus dataset in a specific domain. And by collecting common vocabulary and extracting new words in the domain, we also construct a domain vocabulary. Then, the Chinese word boundary information is added to the Transformer-XL language model to make the model can better adapt to the characteristics of the domain. Finally, our experimental results show that the method is effective on the dataset we provided. It can further reduce the Character Error Rate (CER) in speech recognition.