Xuanang Zheng, Lingli Zhang, Wei Zheng, Wen-zhong Hu
{"title":"基于跨距的主客体联合提取及多头自注意关系分类","authors":"Xuanang Zheng, Lingli Zhang, Wei Zheng, Wen-zhong Hu","doi":"10.1109/icicn52636.2021.9673905","DOIUrl":null,"url":null,"abstract":"Entity recognition and relation extraction are two central tasks for extracting information from unstructured text. The current popular models are to perform joint entity and relationship extraction at the span level. However, they do not take full advantage of the uneven distribution of information values in natural languages and their essence of extracting entities before classifying relations leads to entity redundancy. To solve these problems, we propose a novel span-based joint entity and relationship extraction model. We embed the multi-head self-attention layer to the existing span-based entity and relation joint extraction model to leverage valuable information such as semantic and syntactic features in the input sentences. We also propose a novel method to extract subject and object, which can alleviate the entity redundancy problem. In the ablation experiments, we demonstrate the benefits of both improvements for extracting relation triples. The micro f1 of our model on Con1104 is 72.82, the macro f1 is 74.20, and the f1 on SciERC is 52.03, which are better than the currently widely used model.","PeriodicalId":231379,"journal":{"name":"2021 IEEE 9th International Conference on Information, Communication and Networks (ICICN)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Span-based Joint Extracting Subjects and Objects and Classifying Relations with Multi-head Self-attention\",\"authors\":\"Xuanang Zheng, Lingli Zhang, Wei Zheng, Wen-zhong Hu\",\"doi\":\"10.1109/icicn52636.2021.9673905\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Entity recognition and relation extraction are two central tasks for extracting information from unstructured text. The current popular models are to perform joint entity and relationship extraction at the span level. However, they do not take full advantage of the uneven distribution of information values in natural languages and their essence of extracting entities before classifying relations leads to entity redundancy. To solve these problems, we propose a novel span-based joint entity and relationship extraction model. We embed the multi-head self-attention layer to the existing span-based entity and relation joint extraction model to leverage valuable information such as semantic and syntactic features in the input sentences. We also propose a novel method to extract subject and object, which can alleviate the entity redundancy problem. In the ablation experiments, we demonstrate the benefits of both improvements for extracting relation triples. The micro f1 of our model on Con1104 is 72.82, the macro f1 is 74.20, and the f1 on SciERC is 52.03, which are better than the currently widely used model.\",\"PeriodicalId\":231379,\"journal\":{\"name\":\"2021 IEEE 9th International Conference on Information, Communication and Networks (ICICN)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 9th International Conference on Information, Communication and Networks (ICICN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/icicn52636.2021.9673905\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 9th International Conference on Information, Communication and Networks (ICICN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icicn52636.2021.9673905","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Span-based Joint Extracting Subjects and Objects and Classifying Relations with Multi-head Self-attention
Entity recognition and relation extraction are two central tasks for extracting information from unstructured text. The current popular models are to perform joint entity and relationship extraction at the span level. However, they do not take full advantage of the uneven distribution of information values in natural languages and their essence of extracting entities before classifying relations leads to entity redundancy. To solve these problems, we propose a novel span-based joint entity and relationship extraction model. We embed the multi-head self-attention layer to the existing span-based entity and relation joint extraction model to leverage valuable information such as semantic and syntactic features in the input sentences. We also propose a novel method to extract subject and object, which can alleviate the entity redundancy problem. In the ablation experiments, we demonstrate the benefits of both improvements for extracting relation triples. The micro f1 of our model on Con1104 is 72.82, the macro f1 is 74.20, and the f1 on SciERC is 52.03, which are better than the currently widely used model.