基于Bi-GRU的中文实体关系抽取方法

Jian-qiong Xiao, Zhi-yong Zhou, Xingrong Luo
{"title":"基于Bi-GRU的中文实体关系抽取方法","authors":"Jian-qiong Xiao, Zhi-yong Zhou, Xingrong Luo","doi":"10.2991/MASTA-19.2019.9","DOIUrl":null,"url":null,"abstract":"In order to solve some defects of single deep neural network in Chinese entity Relationship Extraction task, a hybrid neural network entity relationship extraction model is designed and implemented in this paper. The model combines convolution network and bidirectional GRU model with a unified architecture, by defining varisized regional list embedding, it produces nonobjective feature representations of word vectors in distinction positions, and it has only Chinese character vectors and Chinese character word vectors, without position embedding. The laboratory findings show that our method is very effective on the Chinese corpus ACE2005 dataset about entities extraction task. Introduction Aiming at the problem that the traditional CNN model ignores the text context and leads to the lack of text semantics, a convolution layer improvement algorithm is proposed in this paper. By stripping the convolution layer from the CNN model, the convolution layer structure is improved, by defining a varisized sizes regional list embedding, and producing nonobjective feature representation of word vectors at imparity locations, which results in more accurate feature representation by fusing multiple local features, and position vectors aren’t required. Related Work Many researchers have putted forward to a number of solid relational extraction methods based on deep neural networks. Liu et al. [1] is first team who used convolution neural network to automatically learn sentence representation for relational classification tasks, the characteristics of lexical features, lexicality and so on are added to the model, their model get F1 value in the corpus ACE2005 dataset, and exceed the kernel function method 9%. Dong-xu Zhang et al. [2] used RNN to get varisized location feature by training corpus. Zhang [3] proposed using bidirectional long short-term memory network(Bi-LSTM) to build whole sentences extract model, this model used many features, include NLP tools and lexical resources, POS, NER and so on, of course, it achieved the state-of-the-art result. In order to make full use of the competitive advantages of existing neural networks, Sun ziyang et al. [4] also used BiLSTM to model sentence dependency shortest path, and this model use out of CNN as input of LSTM to train the model. This method took full comprehensive advantage of Bi-LSTM model and the CNN model, because Bi-LSTM is good at capturing distance mutual relationship, while CNN can capture the local flat characteristics of sentences. In view of above research, we propose a method for Chinese entity relationship extraction based on bidirectional Gated Recurrent Unit (RLE-BiGRU), it is a new means by defining a new vector (RLE). Experiments result is F1-score of 85.3% on the Chinese corpus ACE2005 dataset, which shows the validity of the work in this paper for the Chinese entity relations extraction. Model In this section we introduce our RLE-BiGRU model in detail. Figure 1 is the model photograph. Our model has six constituent parts: International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019) Copyright © 2019, the Authors. Published by Atlantis Press. This is an open access article under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/). Advances in Intelligent Systems Research, volume 168","PeriodicalId":103896,"journal":{"name":"Proceedings of the 2019 International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A Method for Chinese Entity Relationship Extraction Based on Bi-GRU\",\"authors\":\"Jian-qiong Xiao, Zhi-yong Zhou, Xingrong Luo\",\"doi\":\"10.2991/MASTA-19.2019.9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In order to solve some defects of single deep neural network in Chinese entity Relationship Extraction task, a hybrid neural network entity relationship extraction model is designed and implemented in this paper. The model combines convolution network and bidirectional GRU model with a unified architecture, by defining varisized regional list embedding, it produces nonobjective feature representations of word vectors in distinction positions, and it has only Chinese character vectors and Chinese character word vectors, without position embedding. The laboratory findings show that our method is very effective on the Chinese corpus ACE2005 dataset about entities extraction task. Introduction Aiming at the problem that the traditional CNN model ignores the text context and leads to the lack of text semantics, a convolution layer improvement algorithm is proposed in this paper. By stripping the convolution layer from the CNN model, the convolution layer structure is improved, by defining a varisized sizes regional list embedding, and producing nonobjective feature representation of word vectors at imparity locations, which results in more accurate feature representation by fusing multiple local features, and position vectors aren’t required. Related Work Many researchers have putted forward to a number of solid relational extraction methods based on deep neural networks. Liu et al. [1] is first team who used convolution neural network to automatically learn sentence representation for relational classification tasks, the characteristics of lexical features, lexicality and so on are added to the model, their model get F1 value in the corpus ACE2005 dataset, and exceed the kernel function method 9%. Dong-xu Zhang et al. [2] used RNN to get varisized location feature by training corpus. Zhang [3] proposed using bidirectional long short-term memory network(Bi-LSTM) to build whole sentences extract model, this model used many features, include NLP tools and lexical resources, POS, NER and so on, of course, it achieved the state-of-the-art result. In order to make full use of the competitive advantages of existing neural networks, Sun ziyang et al. [4] also used BiLSTM to model sentence dependency shortest path, and this model use out of CNN as input of LSTM to train the model. This method took full comprehensive advantage of Bi-LSTM model and the CNN model, because Bi-LSTM is good at capturing distance mutual relationship, while CNN can capture the local flat characteristics of sentences. In view of above research, we propose a method for Chinese entity relationship extraction based on bidirectional Gated Recurrent Unit (RLE-BiGRU), it is a new means by defining a new vector (RLE). Experiments result is F1-score of 85.3% on the Chinese corpus ACE2005 dataset, which shows the validity of the work in this paper for the Chinese entity relations extraction. Model In this section we introduce our RLE-BiGRU model in detail. Figure 1 is the model photograph. Our model has six constituent parts: International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019) Copyright © 2019, the Authors. Published by Atlantis Press. This is an open access article under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/). Advances in Intelligent Systems Research, volume 168\",\"PeriodicalId\":103896,\"journal\":{\"name\":\"Proceedings of the 2019 International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019)\",\"volume\":\"17 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2019 International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2991/MASTA-19.2019.9\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2019 International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2991/MASTA-19.2019.9","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

为了解决单一深度神经网络在中文实体关系抽取任务中的一些缺陷,设计并实现了一种混合神经网络实体关系抽取模型。该模型将卷积网络与双向GRU模型结合在一起,采用统一的体系结构,通过定义多样化的区域表嵌入,生成词向量在不同位置的非客观特征表示,并且只有汉字向量和汉字词向量,没有位置嵌入。实验结果表明,该方法在中文语料库ACE2005数据集的实体抽取任务上是非常有效的。针对传统CNN模型忽略文本上下文导致文本语义缺失的问题,本文提出了一种卷积层改进算法。通过从CNN模型中剥离卷积层,改进卷积层结构,通过定义不同大小的区域列表嵌入,并在不相等位置产生词向量的非客观特征表示,从而通过融合多个局部特征获得更准确的特征表示,并且不需要位置向量。许多研究者提出了许多基于深度神经网络的实体关系提取方法。Liu等[1]是第一个利用卷积神经网络自动学习关系分类任务的句子表示的团队,将词法特征、词性等特征加入到模型中,其模型在语料库ACE2005数据集中获得F1值,超过核函数方法9%。张东旭等[2]利用RNN通过训练语料库获得多样化的位置特征。Zhang[3]提出使用双向长短期记忆网络(Bi-LSTM)构建整句提取模型,该模型使用了许多特征,包括NLP工具和词汇资源、POS、NER等,当然也取得了最先进的结果。该方法充分综合了Bi-LSTM模型和CNN模型的优点,因为Bi-LSTM善于捕捉距离相互关系,而CNN则能捕捉句子的局部平坦特征。针对上述研究,本文提出了一种基于双向门控循环单元(RLE- bigru)的中文实体关系提取方法,该方法通过定义一个新的向量(RLE)来实现。在中文语料库ACE2005数据集上的实验结果为f1得分85.3%,表明本文的工作对中文实体关系抽取是有效的。在本节中我们将详细介绍我们的RLE-BiGRU模型。图1是模型照片。我们的模型有六个组成部分:国际建模、分析、仿真技术与应用会议(MASTA 2019)版权所有©2019,作者。亚特兰蒂斯出版社出版。这是一篇基于CC BY-NC许可(http://creativecommons.org/licenses/by-nc/4.0/)的开放获取文章。智能系统研究进展,第168卷
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Method for Chinese Entity Relationship Extraction Based on Bi-GRU
In order to solve some defects of single deep neural network in Chinese entity Relationship Extraction task, a hybrid neural network entity relationship extraction model is designed and implemented in this paper. The model combines convolution network and bidirectional GRU model with a unified architecture, by defining varisized regional list embedding, it produces nonobjective feature representations of word vectors in distinction positions, and it has only Chinese character vectors and Chinese character word vectors, without position embedding. The laboratory findings show that our method is very effective on the Chinese corpus ACE2005 dataset about entities extraction task. Introduction Aiming at the problem that the traditional CNN model ignores the text context and leads to the lack of text semantics, a convolution layer improvement algorithm is proposed in this paper. By stripping the convolution layer from the CNN model, the convolution layer structure is improved, by defining a varisized sizes regional list embedding, and producing nonobjective feature representation of word vectors at imparity locations, which results in more accurate feature representation by fusing multiple local features, and position vectors aren’t required. Related Work Many researchers have putted forward to a number of solid relational extraction methods based on deep neural networks. Liu et al. [1] is first team who used convolution neural network to automatically learn sentence representation for relational classification tasks, the characteristics of lexical features, lexicality and so on are added to the model, their model get F1 value in the corpus ACE2005 dataset, and exceed the kernel function method 9%. Dong-xu Zhang et al. [2] used RNN to get varisized location feature by training corpus. Zhang [3] proposed using bidirectional long short-term memory network(Bi-LSTM) to build whole sentences extract model, this model used many features, include NLP tools and lexical resources, POS, NER and so on, of course, it achieved the state-of-the-art result. In order to make full use of the competitive advantages of existing neural networks, Sun ziyang et al. [4] also used BiLSTM to model sentence dependency shortest path, and this model use out of CNN as input of LSTM to train the model. This method took full comprehensive advantage of Bi-LSTM model and the CNN model, because Bi-LSTM is good at capturing distance mutual relationship, while CNN can capture the local flat characteristics of sentences. In view of above research, we propose a method for Chinese entity relationship extraction based on bidirectional Gated Recurrent Unit (RLE-BiGRU), it is a new means by defining a new vector (RLE). Experiments result is F1-score of 85.3% on the Chinese corpus ACE2005 dataset, which shows the validity of the work in this paper for the Chinese entity relations extraction. Model In this section we introduce our RLE-BiGRU model in detail. Figure 1 is the model photograph. Our model has six constituent parts: International Conference on Modeling, Analysis, Simulation Technologies and Applications (MASTA 2019) Copyright © 2019, the Authors. Published by Atlantis Press. This is an open access article under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/). Advances in Intelligent Systems Research, volume 168
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信