An Extensive Study on Pretrained Models for Natural Language Processing Based on Transformers

M. Ramprasath, K. Dhanasekaran, T. Karthick, R. Velumani, P. Sudhakaran
{"title":"An Extensive Study on Pretrained Models for Natural Language Processing Based on Transformers","authors":"M. Ramprasath, K. Dhanasekaran, T. Karthick, R. Velumani, P. Sudhakaran","doi":"10.1109/ICEARS53579.2022.9752241","DOIUrl":null,"url":null,"abstract":"In recent years, Pretraining Language Models based on Transformers -Natural Language Processing (PLMT-NLP) have been highly successful in nearly every NLP task. In the beginning, Generative Pre-trained model-based Transformer, BERT- Bidirectional Encoder model Representations using Transformers was used to develop these models. Models constructed on transformers, Self-supervise knowledge acquiring, and transfer learning establish the foundation of these designs. Transformed-based pre-trained models acquire common linguistic illustrations from vast amounts of textual information through self-supervised model and apply this information to downstream tasks. To eliminate the need to retrain downstream models, these models provide a solid foundation of knowledge. In this paper, the enhanced learning on PLMT-NLP has been discussed. Initially, a quick introduction to self-supervised learning is presented, then diverse core concepts used in PLMT-NLP are explained. Furthermore, a list of relevant libraries for working with PLMT-NLP has been provided. Lastly, the paper discusses about the upcoming research directions that will further improve these models. Because of its thoroughness and relevance to current PLMT-NLP developments, this survey study will positively serve as a valuable resource for those seeking to understand both basic ideas and new developments better.","PeriodicalId":252961,"journal":{"name":"2022 International Conference on Electronics and Renewable Systems (ICEARS)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Electronics and Renewable Systems (ICEARS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEARS53579.2022.9752241","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

In recent years, Pretraining Language Models based on Transformers -Natural Language Processing (PLMT-NLP) have been highly successful in nearly every NLP task. In the beginning, Generative Pre-trained model-based Transformer, BERT- Bidirectional Encoder model Representations using Transformers was used to develop these models. Models constructed on transformers, Self-supervise knowledge acquiring, and transfer learning establish the foundation of these designs. Transformed-based pre-trained models acquire common linguistic illustrations from vast amounts of textual information through self-supervised model and apply this information to downstream tasks. To eliminate the need to retrain downstream models, these models provide a solid foundation of knowledge. In this paper, the enhanced learning on PLMT-NLP has been discussed. Initially, a quick introduction to self-supervised learning is presented, then diverse core concepts used in PLMT-NLP are explained. Furthermore, a list of relevant libraries for working with PLMT-NLP has been provided. Lastly, the paper discusses about the upcoming research directions that will further improve these models. Because of its thoroughness and relevance to current PLMT-NLP developments, this survey study will positively serve as a valuable resource for those seeking to understand both basic ideas and new developments better.
基于变形器的自然语言处理预训练模型的广泛研究
近年来,基于变形器-自然语言处理(PLMT-NLP)的预训练语言模型在几乎所有的NLP任务中都取得了巨大的成功。首先,基于生成预训练模型的Transformer, BERT-使用Transformer的双向编码器模型表示用于开发这些模型。基于变压器的模型构建、自我监督知识获取和迁移学习为这些设计奠定了基础。基于转换的预训练模型通过自监督模型从大量文本信息中获取共同的语言插图,并将这些信息应用于下游任务。为了消除对下游模型进行再培训的需要,这些模型提供了坚实的知识基础。本文对PLMT-NLP的强化学习进行了讨论。首先,介绍了自我监督学习,然后解释了PLMT-NLP中使用的各种核心概念。此外,还提供了使用PLMT-NLP的相关库的列表。最后,对未来的研究方向进行了展望,以进一步完善这些模型。由于其彻底性和与当前PLMT-NLP发展的相关性,这项调查研究将积极地为那些寻求更好地理解基本思想和新发展的人提供宝贵的资源。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信