Geonsik Youn, Bohan Yoon, Seungbin Ji, Dahee Ko, J. Rhee
{"title":"基于混合的命名实体识别交叉一致性训练","authors":"Geonsik Youn, Bohan Yoon, Seungbin Ji, Dahee Ko, J. Rhee","doi":"10.1145/3571560.3571576","DOIUrl":null,"url":null,"abstract":"Named Entity Recognition (NER) is one of the first stages in deep natural language understanding. The state-of-the-art deep NER models are dependent on high-quality and massive datasets. Also, the NER tasks require token-level labels. For this reason, there is a problem that annotating many sentences for the NER tasks is time-consuming and expensive. To solve this problem, many prior studies have been conducted to use the auto annotated weakly labeled data. However, the weakly labeled data contains a lot of noises that are obstructive to the training of NER models. We propose to use MixUp and cross-consistency training (CCT) together as a strategy to use weakly labeled data for NER tasks. In this study, the proposed method stems from the idea that MixUp, which was recently considered the data augmentation strategy, hinders the NER model training. Inspired by this point, we propose to use MixUp as a perturbation of cross-consistency training for NER. Experiments conducted on several NER benchmarks demonstrate the proposed method achieves improved performance compared to employing only a few human-annotated data.","PeriodicalId":143909,"journal":{"name":"Proceedings of the 6th International Conference on Advances in Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"MixUp based Cross-Consistency Training for Named Entity Recognition\",\"authors\":\"Geonsik Youn, Bohan Yoon, Seungbin Ji, Dahee Ko, J. Rhee\",\"doi\":\"10.1145/3571560.3571576\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Named Entity Recognition (NER) is one of the first stages in deep natural language understanding. The state-of-the-art deep NER models are dependent on high-quality and massive datasets. Also, the NER tasks require token-level labels. For this reason, there is a problem that annotating many sentences for the NER tasks is time-consuming and expensive. To solve this problem, many prior studies have been conducted to use the auto annotated weakly labeled data. However, the weakly labeled data contains a lot of noises that are obstructive to the training of NER models. We propose to use MixUp and cross-consistency training (CCT) together as a strategy to use weakly labeled data for NER tasks. In this study, the proposed method stems from the idea that MixUp, which was recently considered the data augmentation strategy, hinders the NER model training. Inspired by this point, we propose to use MixUp as a perturbation of cross-consistency training for NER. Experiments conducted on several NER benchmarks demonstrate the proposed method achieves improved performance compared to employing only a few human-annotated data.\",\"PeriodicalId\":143909,\"journal\":{\"name\":\"Proceedings of the 6th International Conference on Advances in Artificial Intelligence\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 6th International Conference on Advances in Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3571560.3571576\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 6th International Conference on Advances in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3571560.3571576","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
MixUp based Cross-Consistency Training for Named Entity Recognition
Named Entity Recognition (NER) is one of the first stages in deep natural language understanding. The state-of-the-art deep NER models are dependent on high-quality and massive datasets. Also, the NER tasks require token-level labels. For this reason, there is a problem that annotating many sentences for the NER tasks is time-consuming and expensive. To solve this problem, many prior studies have been conducted to use the auto annotated weakly labeled data. However, the weakly labeled data contains a lot of noises that are obstructive to the training of NER models. We propose to use MixUp and cross-consistency training (CCT) together as a strategy to use weakly labeled data for NER tasks. In this study, the proposed method stems from the idea that MixUp, which was recently considered the data augmentation strategy, hinders the NER model training. Inspired by this point, we propose to use MixUp as a perturbation of cross-consistency training for NER. Experiments conducted on several NER benchmarks demonstrate the proposed method achieves improved performance compared to employing only a few human-annotated data.