Harnessing high-quality pseudo-labels for robust few-shot nested named entity recognition

IF 7.5 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS
Hong Ming , Jiaoyun Yang , Shuo Liu , Lili Jiang , Ning An
{"title":"Harnessing high-quality pseudo-labels for robust few-shot nested named entity recognition","authors":"Hong Ming ,&nbsp;Jiaoyun Yang ,&nbsp;Shuo Liu ,&nbsp;Lili Jiang ,&nbsp;Ning An","doi":"10.1016/j.engappai.2025.110992","DOIUrl":null,"url":null,"abstract":"<div><div>Few-shot Named Entity Recognition (NER) methods have shown initial effectiveness in flat NER tasks. However, these methods often prioritize optimizing models with a small annotated support set, neglecting the high-quality data within the unlabeled query set. Furthermore, existing few-shot NER models struggle with nested entity challenges due to linguistic or structural complexities. In this study, we introduce <strong>R</strong>etrieving h<strong>i</strong>gh-quality pseudo-label <strong>T</strong>uning, RiTNER, a framework designed to address few-shot nested named entity recognition tasks by leveraging high-quality data from the query set. RiTNER comprises two main components: (1) contrastive span classification, which clusters entities into corresponding prototypes and generates high-quality pseudo-labels from the unlabeled data, and (2) masked pseudo-data tuning, which generates a masked pseudo dataset and then uses it to optimize the model and enhance span classification. We train RiTNER on an English dataset and evaluate it on both English nested datasets and cross-lingual nested datasets. The results show that RiTNER outperforms the top-performing baseline models by 1.67%, and 3.04% in the English 5-shot task, as well as the cross-lingual 5-shot tasks, respectively.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"156 ","pages":"Article 110992"},"PeriodicalIF":7.5000,"publicationDate":"2025-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625009923","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Few-shot Named Entity Recognition (NER) methods have shown initial effectiveness in flat NER tasks. However, these methods often prioritize optimizing models with a small annotated support set, neglecting the high-quality data within the unlabeled query set. Furthermore, existing few-shot NER models struggle with nested entity challenges due to linguistic or structural complexities. In this study, we introduce Retrieving high-quality pseudo-label Tuning, RiTNER, a framework designed to address few-shot nested named entity recognition tasks by leveraging high-quality data from the query set. RiTNER comprises two main components: (1) contrastive span classification, which clusters entities into corresponding prototypes and generates high-quality pseudo-labels from the unlabeled data, and (2) masked pseudo-data tuning, which generates a masked pseudo dataset and then uses it to optimize the model and enhance span classification. We train RiTNER on an English dataset and evaluate it on both English nested datasets and cross-lingual nested datasets. The results show that RiTNER outperforms the top-performing baseline models by 1.67%, and 3.04% in the English 5-shot task, as well as the cross-lingual 5-shot tasks, respectively.
利用高质量的伪标签进行鲁棒的少镜头嵌套命名实体识别
在平面命名实体识别(NER)任务中,少弹命名实体识别(NER)方法已经显示出初步的有效性。然而,这些方法通常使用一个小的带注释的支持集来优先优化模型,而忽略了未标记查询集中的高质量数据。此外,由于语言或结构的复杂性,现有的少镜头NER模型面临嵌套实体的挑战。在本研究中,我们介绍了检索高质量伪标签调优,RiTNER,这是一个框架,旨在通过利用来自查询集的高质量数据来解决少量嵌套命名实体识别任务。RiTNER包括两个主要组成部分:(1)对比跨度分类,将实体聚类到相应的原型中,并从未标记的数据中生成高质量的伪标签;(2)掩码伪数据调优,生成掩码伪数据集,然后使用它来优化模型并增强跨度分类。我们在英语数据集上训练RiTNER,并在英语嵌套数据集和跨语言嵌套数据集上对其进行评估。结果表明,在英语5-shot任务和跨语言5-shot任务中,RiTNER分别比表现最好的基线模型高出1.67%和3.04%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Engineering Applications of Artificial Intelligence
Engineering Applications of Artificial Intelligence 工程技术-工程:电子与电气
CiteScore
9.60
自引率
10.00%
发文量
505
审稿时长
68 days
期刊介绍: Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信