端到端RNN中MTL的浅学习,用于基本序列标注

Rajat Subhra Bhowmick, Trina Ghosh, Astha Singh, Sayak Chakraborty, J. Sil
{"title":"端到端RNN中MTL的浅学习,用于基本序列标注","authors":"Rajat Subhra Bhowmick, Trina Ghosh, Astha Singh, Sayak Chakraborty, J. Sil","doi":"10.1145/3474124.3474160","DOIUrl":null,"url":null,"abstract":"Multitask learning (MTL) has been successfully applied for sequence tagging tasks in many machine learning approaches. The idea of MTL is employed to obtain better performance by sharing information among the tasks compared to a single task. However, most of the MTL in end-to-end sequence tagging models perform by sharing the entire set of parameters resulting in overwriting of parameters and need retraining of all the parameters. In the paper, a novel pipeline architecture has been proposed that effectively combines two RNN based sub-networks for sequence tagging tasks with minimal training overhead and truly acts as an end-to-end system. The proposed architecture performs Named entity recognition (NER) tagging as the primary sequence tagging task along with phrase tagging that works as the assistance sequence tagging task. To utilize the learning from the assisted network, we modify the base network by adding fully connected (FC) layers for NER. Our MTL approach tries to retain the learning parameters from individual tasks and, retraining is done only to the FC(shallow) layers of the modified base network. To validate the proposed MTL settings, we train on CoNLL 2003 corpus and compare the result with previously well established end-to-end based NER tagging models.","PeriodicalId":144611,"journal":{"name":"2021 Thirteenth International Conference on Contemporary Computing (IC3-2021)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Shallow learning for MTL in end-to-end RNN for basic sequence tagging\",\"authors\":\"Rajat Subhra Bhowmick, Trina Ghosh, Astha Singh, Sayak Chakraborty, J. Sil\",\"doi\":\"10.1145/3474124.3474160\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multitask learning (MTL) has been successfully applied for sequence tagging tasks in many machine learning approaches. The idea of MTL is employed to obtain better performance by sharing information among the tasks compared to a single task. However, most of the MTL in end-to-end sequence tagging models perform by sharing the entire set of parameters resulting in overwriting of parameters and need retraining of all the parameters. In the paper, a novel pipeline architecture has been proposed that effectively combines two RNN based sub-networks for sequence tagging tasks with minimal training overhead and truly acts as an end-to-end system. The proposed architecture performs Named entity recognition (NER) tagging as the primary sequence tagging task along with phrase tagging that works as the assistance sequence tagging task. To utilize the learning from the assisted network, we modify the base network by adding fully connected (FC) layers for NER. Our MTL approach tries to retain the learning parameters from individual tasks and, retraining is done only to the FC(shallow) layers of the modified base network. To validate the proposed MTL settings, we train on CoNLL 2003 corpus and compare the result with previously well established end-to-end based NER tagging models.\",\"PeriodicalId\":144611,\"journal\":{\"name\":\"2021 Thirteenth International Conference on Contemporary Computing (IC3-2021)\",\"volume\":\"9 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-08-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 Thirteenth International Conference on Contemporary Computing (IC3-2021)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3474124.3474160\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 Thirteenth International Conference on Contemporary Computing (IC3-2021)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3474124.3474160","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

在许多机器学习方法中,多任务学习(MTL)已经成功地应用于序列标记任务。与单个任务相比,MTL的思想通过在任务之间共享信息来获得更好的性能。然而,端到端序列标记模型中的大多数MTL是通过共享整个参数集来执行的,这导致了参数的覆盖,并且需要对所有参数进行重新训练。本文提出了一种新的管道架构,该架构有效地结合了两个基于RNN的子网络,以最小的训练开销完成序列标记任务,并真正充当端到端系统。所提出的体系结构将命名实体识别(NER)标记作为主要序列标记任务,并将短语标记作为辅助序列标记任务。为了利用辅助网络的学习,我们通过为NER添加全连接(FC)层来修改基础网络。我们的MTL方法试图保留来自单个任务的学习参数,并且只对修改后的基本网络的FC(浅)层进行再训练。为了验证提出的MTL设置,我们在CoNLL 2003语料库上进行训练,并将结果与先前建立的基于端到端的NER标记模型进行比较。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Shallow learning for MTL in end-to-end RNN for basic sequence tagging
Multitask learning (MTL) has been successfully applied for sequence tagging tasks in many machine learning approaches. The idea of MTL is employed to obtain better performance by sharing information among the tasks compared to a single task. However, most of the MTL in end-to-end sequence tagging models perform by sharing the entire set of parameters resulting in overwriting of parameters and need retraining of all the parameters. In the paper, a novel pipeline architecture has been proposed that effectively combines two RNN based sub-networks for sequence tagging tasks with minimal training overhead and truly acts as an end-to-end system. The proposed architecture performs Named entity recognition (NER) tagging as the primary sequence tagging task along with phrase tagging that works as the assistance sequence tagging task. To utilize the learning from the assisted network, we modify the base network by adding fully connected (FC) layers for NER. Our MTL approach tries to retain the learning parameters from individual tasks and, retraining is done only to the FC(shallow) layers of the modified base network. To validate the proposed MTL settings, we train on CoNLL 2003 corpus and compare the result with previously well established end-to-end based NER tagging models.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信