DFS-NER: Description Enhanced Few-shot NER via Prompt Learning and Meta-Learning

Huinan Huang, Yuming Feng, Xiaolong Jin, Saiping Guan, Jiafeng Guo
{"title":"DFS-NER: Description Enhanced Few-shot NER via Prompt Learning and Meta-Learning","authors":"Huinan Huang, Yuming Feng, Xiaolong Jin, Saiping Guan, Jiafeng Guo","doi":"10.1109/WI-IAT55865.2022.00131","DOIUrl":null,"url":null,"abstract":"Named Entity Recognition (NER) is a very common task in many social good related domains. Recently, deep learning based NER has gradually matured, but still faces the scarcity problem of labeled data in specific domains. Therefore, researchers focus on few-shot NER to reduce the model’s data dependence and enhance the transferability of the model. However, existing works usually cannot adapt to new entity types and are prone to the so-called negative transfer problem. Therefore, in this paper we propose a type- Description-enhanced Few Shot NER model, called DFS-NER, which effectively integrates the prompt learning paradigm and the meta-learning framework. DFS-NER performs well under frozen pre-training model parameters through continuous templates. It realizes efficient source domain training and target domain parameter fine-tuning through the metalearning framework. We enhance the robustness of entity- type prototype representations by introducing word-word- level and word-type-level contrastive learning objectives and capsule networks as the induction module. Simultaneously, based on discrete prompt learning, a masked-language model learning objective guided by type description is proposed, which can well absorb the semantic information of entity types. Experiments on commonly used datasets, including, SNIPS, Few-NERD, and MIT Movie show that DFS-NER basically surpasses baseline models and achieves the state-of-the-art performance.","PeriodicalId":345445,"journal":{"name":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","volume":"143 8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WI-IAT55865.2022.00131","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Named Entity Recognition (NER) is a very common task in many social good related domains. Recently, deep learning based NER has gradually matured, but still faces the scarcity problem of labeled data in specific domains. Therefore, researchers focus on few-shot NER to reduce the model’s data dependence and enhance the transferability of the model. However, existing works usually cannot adapt to new entity types and are prone to the so-called negative transfer problem. Therefore, in this paper we propose a type- Description-enhanced Few Shot NER model, called DFS-NER, which effectively integrates the prompt learning paradigm and the meta-learning framework. DFS-NER performs well under frozen pre-training model parameters through continuous templates. It realizes efficient source domain training and target domain parameter fine-tuning through the metalearning framework. We enhance the robustness of entity- type prototype representations by introducing word-word- level and word-type-level contrastive learning objectives and capsule networks as the induction module. Simultaneously, based on discrete prompt learning, a masked-language model learning objective guided by type description is proposed, which can well absorb the semantic information of entity types. Experiments on commonly used datasets, including, SNIPS, Few-NERD, and MIT Movie show that DFS-NER basically surpasses baseline models and achieves the state-of-the-art performance.
DFS-NER:描述通过提示学习和元学习增强的少射NER
命名实体识别(NER)是许多社会公益相关领域中非常常见的任务。近年来,基于深度学习的NER逐渐成熟,但仍然面临着特定领域标记数据的稀缺性问题。因此,为了降低模型对数据的依赖性,增强模型的可移植性,研究人员将重点放在了少射NER上。然而,现有的作品往往不能适应新的实体类型,容易出现所谓的负迁移问题。因此,本文提出了一种类型描述增强的Few Shot NER模型,称为DFS-NER,该模型有效地集成了提示学习范式和元学习框架。DFS-NER通过连续模板在固定的预训练模型参数下表现良好。通过元学习框架实现高效的源域训练和目标域参数微调。我们通过引入词-词级和词-类型级对比学习目标和胶囊网络作为归纳模块来增强实体型原型表示的鲁棒性。同时,在离散提示学习的基础上,提出了一种以类型描述为导向的掩蔽语言模型学习目标,能够很好地吸收实体类型的语义信息。在SNIPS、Few-NERD和MIT Movie等常用数据集上的实验表明,DFS-NER基本超过了基线模型,达到了最先进的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信