A. A. Maksutov, Vladimir I. Zamyatovskiy, Viacheslav O. Morozov, S. Dmitriev
{"title":"词性标注的Transformer神经网络结构","authors":"A. A. Maksutov, Vladimir I. Zamyatovskiy, Viacheslav O. Morozov, S. Dmitriev","doi":"10.1109/ElConRus51938.2021.9396231","DOIUrl":null,"url":null,"abstract":"Part-of-speech tagging (POS tagging) is one of the most important tasks in natural language processing. This process implies determining part of speech and assigning an appropriate tag for each word in given sentence. The resulting tag sequence can be used as is and as a part of more complicated tasks, such as dependency and constituency parsing. This task belongs to sequence-to-sequence tasks and multilayer bidirectional LSTM networks are commonly used for POS tagging. Such networks are rather slow in terms of training and processing large amounts of information due to sequential computation of each timestamp from the input sequence. This paper is focused on developing an accurate model for POS tagging that uses the original Transformer neural network architecture.","PeriodicalId":447345,"journal":{"name":"2021 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"The Transformer Neural Network Architecture for Part-of-Speech Tagging\",\"authors\":\"A. A. Maksutov, Vladimir I. Zamyatovskiy, Viacheslav O. Morozov, S. Dmitriev\",\"doi\":\"10.1109/ElConRus51938.2021.9396231\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Part-of-speech tagging (POS tagging) is one of the most important tasks in natural language processing. This process implies determining part of speech and assigning an appropriate tag for each word in given sentence. The resulting tag sequence can be used as is and as a part of more complicated tasks, such as dependency and constituency parsing. This task belongs to sequence-to-sequence tasks and multilayer bidirectional LSTM networks are commonly used for POS tagging. Such networks are rather slow in terms of training and processing large amounts of information due to sequential computation of each timestamp from the input sequence. This paper is focused on developing an accurate model for POS tagging that uses the original Transformer neural network architecture.\",\"PeriodicalId\":447345,\"journal\":{\"name\":\"2021 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus)\",\"volume\":\"61 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-01-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ElConRus51938.2021.9396231\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ElConRus51938.2021.9396231","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Transformer Neural Network Architecture for Part-of-Speech Tagging
Part-of-speech tagging (POS tagging) is one of the most important tasks in natural language processing. This process implies determining part of speech and assigning an appropriate tag for each word in given sentence. The resulting tag sequence can be used as is and as a part of more complicated tasks, such as dependency and constituency parsing. This task belongs to sequence-to-sequence tasks and multilayer bidirectional LSTM networks are commonly used for POS tagging. Such networks are rather slow in terms of training and processing large amounts of information due to sequential computation of each timestamp from the input sequence. This paper is focused on developing an accurate model for POS tagging that uses the original Transformer neural network architecture.