{"title":"Implements of Transformer in NLP and DKT","authors":"Haotong Gong","doi":"10.1109/AIAM57466.2022.00163","DOIUrl":null,"url":null,"abstract":"Transformer is a strong model proposed by Google team in 2017. It was a huge improvement that it entirely abandons the mechanism of Recurrent Neural Network (RNN) and Convolutional Neural Network (RNN). As a result, soon it became a popular choice in a diversity of scenarios. A typical implement of Transformer is for handling text-like input sequences, such as Natural Language Process (NLP) and Knowledge Tracing (KT). Although Transformer is a strong model, it still has a number of improvements. Some state-of-the-art Deep-Learning-based models (e.g., BERT in [2], SAINT in [3], etc.) are based on Transformer. In this paper, I give some examples of application of Transformer or Transformer-based models and summarize the pros and cons of Transformer.","PeriodicalId":439903,"journal":{"name":"2022 4th International Conference on Artificial Intelligence and Advanced Manufacturing (AIAM)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 4th International Conference on Artificial Intelligence and Advanced Manufacturing (AIAM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AIAM57466.2022.00163","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Transformer is a strong model proposed by Google team in 2017. It was a huge improvement that it entirely abandons the mechanism of Recurrent Neural Network (RNN) and Convolutional Neural Network (RNN). As a result, soon it became a popular choice in a diversity of scenarios. A typical implement of Transformer is for handling text-like input sequences, such as Natural Language Process (NLP) and Knowledge Tracing (KT). Although Transformer is a strong model, it still has a number of improvements. Some state-of-the-art Deep-Learning-based models (e.g., BERT in [2], SAINT in [3], etc.) are based on Transformer. In this paper, I give some examples of application of Transformer or Transformer-based models and summarize the pros and cons of Transformer.