{"title":"DFS-NER:描述通过提示学习和元学习增强的少射NER","authors":"Huinan Huang, Yuming Feng, Xiaolong Jin, Saiping Guan, Jiafeng Guo","doi":"10.1109/WI-IAT55865.2022.00131","DOIUrl":null,"url":null,"abstract":"Named Entity Recognition (NER) is a very common task in many social good related domains. Recently, deep learning based NER has gradually matured, but still faces the scarcity problem of labeled data in specific domains. Therefore, researchers focus on few-shot NER to reduce the model’s data dependence and enhance the transferability of the model. However, existing works usually cannot adapt to new entity types and are prone to the so-called negative transfer problem. Therefore, in this paper we propose a type- Description-enhanced Few Shot NER model, called DFS-NER, which effectively integrates the prompt learning paradigm and the meta-learning framework. DFS-NER performs well under frozen pre-training model parameters through continuous templates. It realizes efficient source domain training and target domain parameter fine-tuning through the metalearning framework. We enhance the robustness of entity- type prototype representations by introducing word-word- level and word-type-level contrastive learning objectives and capsule networks as the induction module. Simultaneously, based on discrete prompt learning, a masked-language model learning objective guided by type description is proposed, which can well absorb the semantic information of entity types. Experiments on commonly used datasets, including, SNIPS, Few-NERD, and MIT Movie show that DFS-NER basically surpasses baseline models and achieves the state-of-the-art performance.","PeriodicalId":345445,"journal":{"name":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","volume":"143 8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"DFS-NER: Description Enhanced Few-shot NER via Prompt Learning and Meta-Learning\",\"authors\":\"Huinan Huang, Yuming Feng, Xiaolong Jin, Saiping Guan, Jiafeng Guo\",\"doi\":\"10.1109/WI-IAT55865.2022.00131\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Named Entity Recognition (NER) is a very common task in many social good related domains. Recently, deep learning based NER has gradually matured, but still faces the scarcity problem of labeled data in specific domains. Therefore, researchers focus on few-shot NER to reduce the model’s data dependence and enhance the transferability of the model. However, existing works usually cannot adapt to new entity types and are prone to the so-called negative transfer problem. Therefore, in this paper we propose a type- Description-enhanced Few Shot NER model, called DFS-NER, which effectively integrates the prompt learning paradigm and the meta-learning framework. DFS-NER performs well under frozen pre-training model parameters through continuous templates. It realizes efficient source domain training and target domain parameter fine-tuning through the metalearning framework. We enhance the robustness of entity- type prototype representations by introducing word-word- level and word-type-level contrastive learning objectives and capsule networks as the induction module. Simultaneously, based on discrete prompt learning, a masked-language model learning objective guided by type description is proposed, which can well absorb the semantic information of entity types. Experiments on commonly used datasets, including, SNIPS, Few-NERD, and MIT Movie show that DFS-NER basically surpasses baseline models and achieves the state-of-the-art performance.\",\"PeriodicalId\":345445,\"journal\":{\"name\":\"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)\",\"volume\":\"143 8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WI-IAT55865.2022.00131\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WI-IAT55865.2022.00131","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
DFS-NER: Description Enhanced Few-shot NER via Prompt Learning and Meta-Learning
Named Entity Recognition (NER) is a very common task in many social good related domains. Recently, deep learning based NER has gradually matured, but still faces the scarcity problem of labeled data in specific domains. Therefore, researchers focus on few-shot NER to reduce the model’s data dependence and enhance the transferability of the model. However, existing works usually cannot adapt to new entity types and are prone to the so-called negative transfer problem. Therefore, in this paper we propose a type- Description-enhanced Few Shot NER model, called DFS-NER, which effectively integrates the prompt learning paradigm and the meta-learning framework. DFS-NER performs well under frozen pre-training model parameters through continuous templates. It realizes efficient source domain training and target domain parameter fine-tuning through the metalearning framework. We enhance the robustness of entity- type prototype representations by introducing word-word- level and word-type-level contrastive learning objectives and capsule networks as the induction module. Simultaneously, based on discrete prompt learning, a masked-language model learning objective guided by type description is proposed, which can well absorb the semantic information of entity types. Experiments on commonly used datasets, including, SNIPS, Few-NERD, and MIT Movie show that DFS-NER basically surpasses baseline models and achieves the state-of-the-art performance.