FastSpanNER: Speeding up SpanNER by Named Entity Head Prediction

Q3 Arts and Humanities
Icon Pub Date : 2023-03-01 DOI:10.1109/ICNLP58431.2023.00042
Min Zhang, Yanqing Zhao, Xiaosong Qiao, Song Peng, Shimin Tao, Hao Yang, Ying Qin, Yanfei Jiang
{"title":"FastSpanNER: Speeding up SpanNER by Named Entity Head Prediction","authors":"Min Zhang, Yanqing Zhao, Xiaosong Qiao, Song Peng, Shimin Tao, Hao Yang, Ying Qin, Yanfei Jiang","doi":"10.1109/ICNLP58431.2023.00042","DOIUrl":null,"url":null,"abstract":"Named Entity Recognition (NER) is one of the most fundamental tasks in natural language processing (NLP). Different from the widely-used sequence labeling framework in NER, span prediction based methods are more naturally suitable for the nested NER problem and have received a lot of attention recently. However, classifying the samples generated by traversing all sub-sequences is computational expensive during training and very ineffective at inference. In this paper, we propose the FastSpanNER approach to reduce the computation of both training and inferring. We introduce a task of Named Entity Head (NEH) prediction for each word in given sequence, and perform multi-task learning together with the task of span classification, which uses no more than half of the samples in SpanNER. In the inference phase, only the words predicted as NEHs are used to generate candidate spans for named entity classification. Experimental results on the four standard benchmark datasets (CoNLL2003, MSRA, CNERTA and GENIA) show that our FastSpanNER method not only greatly reduces the computation of training and inferring but also achieves better F1 scores compared with the SpanNER method.","PeriodicalId":53637,"journal":{"name":"Icon","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNLP58431.2023.00042","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
引用次数: 0

Abstract

Named Entity Recognition (NER) is one of the most fundamental tasks in natural language processing (NLP). Different from the widely-used sequence labeling framework in NER, span prediction based methods are more naturally suitable for the nested NER problem and have received a lot of attention recently. However, classifying the samples generated by traversing all sub-sequences is computational expensive during training and very ineffective at inference. In this paper, we propose the FastSpanNER approach to reduce the computation of both training and inferring. We introduce a task of Named Entity Head (NEH) prediction for each word in given sequence, and perform multi-task learning together with the task of span classification, which uses no more than half of the samples in SpanNER. In the inference phase, only the words predicted as NEHs are used to generate candidate spans for named entity classification. Experimental results on the four standard benchmark datasets (CoNLL2003, MSRA, CNERTA and GENIA) show that our FastSpanNER method not only greatly reduces the computation of training and inferring but also achieves better F1 scores compared with the SpanNER method.
FastSpanNER:通过命名实体头部预测加速扳手
命名实体识别(NER)是自然语言处理(NLP)中最基本的任务之一。与NER中广泛使用的序列标记框架不同,基于跨度预测的方法更自然地适用于嵌套NER问题,近年来受到了广泛的关注。然而,通过遍历所有子序列生成的样本进行分类,在训练过程中计算成本很高,在推理时效率非常低。在本文中,我们提出了FastSpanNER方法来减少训练和推断的计算。我们对给定序列中的每个单词引入命名实体头(NEH)预测任务,并结合跨度分类任务进行多任务学习,该任务使用的样本不超过SpanNER的一半。在推理阶段,只有被预测为neh的单词才会被用来为命名实体分类生成候选范围。在四个标准基准数据集(CoNLL2003、MSRA、CNERTA和GENIA)上的实验结果表明,FastSpanNER方法不仅大大减少了训练和推断的计算量,而且与SpanNER方法相比,获得了更好的F1分数。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Icon
Icon Arts and Humanities-History and Philosophy of Science
CiteScore
0.30
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信