Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)最新文献

筛选
英文 中文
Explanation Regeneration via Multi-Hop ILP Inference over Knowledge Base 基于知识库的多跳ILP推理解释再生
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.13
Aayushee Gupta, G. Srinivasaraghavan
{"title":"Explanation Regeneration via Multi-Hop ILP Inference over Knowledge Base","authors":"Aayushee Gupta, G. Srinivasaraghavan","doi":"10.18653/v1/2020.textgraphs-1.13","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.13","url":null,"abstract":"Textgraphs 2020 Workshop organized a shared task on ‘Explanation Regeneration’ that required reconstructing gold explanations for elementary science questions. This work describes our submission to the task which is based on multiple components: a BERT baseline ranking, an Integer Linear Program (ILP) based re-scoring and a regression model for re-ranking the explanation facts. Our system achieved a Mean Average Precision score of 0.3659.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114569263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Graph-based Aspect Representation Learning for Entity Resolution 面向实体解析的基于图的方面表示学习
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.2
Zhenqiang Zhao, Yuchen Guo, Dingxian Wang, Yufang Huang, Xiangnan He, Bin Gu
{"title":"Graph-based Aspect Representation Learning for Entity Resolution","authors":"Zhenqiang Zhao, Yuchen Guo, Dingxian Wang, Yufang Huang, Xiangnan He, Bin Gu","doi":"10.18653/v1/2020.textgraphs-1.2","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.2","url":null,"abstract":"Entity Resolution (ER) identifies records that refer to the same real-world entity. Deep learning approaches improved the generalization ability of entity matching models, but hardly overcame the impact of noisy or incomplete data sources. In real scenes, an entity usually consists of multiple semantic facets, called aspects. In this paper, we focus on entity augmentation, namely retrieving the values of missing aspects. The relationship between aspects is naturally suitable to be represented by a knowledge graph, where entity augmentation can be modeled as a link prediction problem. Our paper proposes a novel graph-based approach to solve entity augmentation. Specifically, we apply a dedicated random walk algorithm, which uses node types to limit the traversal length, and encodes graph structure into low-dimensional embeddings. Thus, the missing aspects could be retrieved by a link prediction model. Furthermore, the augmented aspects with fixed orders are served as the input of a deep Siamese BiLSTM network for entity matching. We compared our method with state-of-the-art methods through extensive experiments on downstream ER tasks. According to the experiment results, our model outperforms other methods on evaluation metrics (accuracy, precision, recall, and f1-score) to a large extent, which demonstrates the effectiveness of our method.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"231 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114656318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-supervised Word Sense Disambiguation Using Example Similarity Graph 基于实例相似图的半监督词义消歧
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.6
Rie Yatabe, Minoru Sasaki
{"title":"Semi-supervised Word Sense Disambiguation Using Example Similarity Graph","authors":"Rie Yatabe, Minoru Sasaki","doi":"10.18653/v1/2020.textgraphs-1.6","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.6","url":null,"abstract":"Word Sense Disambiguation (WSD) is a well-known problem in the natural language processing. In recent years, there has been increasing interest in applying neural net-works and machine learning techniques to solve WSD problems. However, these previ-ous supervised approaches often suffer from the lack of manually sense-tagged exam-ples. In this paper, to solve these problems, we propose a semi-supervised WSD method using graph embeddings based learning method in order to make effective use of labeled and unlabeled examples. The results of the experiments show that the proposed method performs better than the previous semi-supervised WSD method. Moreover, the graph structure between examples is effective for WSD and it is effective to utilize a graph structure obtained by fine-tuning BERT in the proposed method.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132763492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
TextGraphs 2020 Shared Task on Multi-Hop Inference for Explanation Regeneration TextGraphs 2020多跳推理解释再生共享任务
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.10
Peter Alexander Jansen, Dmitry Ustalov
{"title":"TextGraphs 2020 Shared Task on Multi-Hop Inference for Explanation Regeneration","authors":"Peter Alexander Jansen, Dmitry Ustalov","doi":"10.18653/v1/2020.textgraphs-1.10","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.10","url":null,"abstract":"The 2020 Shared Task on Multi-Hop Inference for Explanation Regeneration tasks participants with regenerating large detailed multi-fact explanations for standardized science exam questions. Given a question, correct answer, and knowledge base, models must rank each fact in the knowledge base such that facts most likely to appear in the explanation are ranked highest. Explanations consist of an average of 6 (and as many as 16) facts that span both core scientific knowledge and world knowledge, and form an explicit lexically-connected “explanation graph” describing how the facts interrelate. In this second iteration of the explanation regeneration shared task, participants are supplied with more than double the training and evaluation data of the first shared task, as well as a knowledge base nearly double in size, both of which expand into more challenging scientific topics that increase the difficulty of the task. In total 10 teams participated, and 5 teams submitted system description papers. The best-performing teams significantly increased state-of-the-art performance both in terms of ranking (mean average precision) and inference speed on this challenge task.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115604932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
ChiSquareX at TextGraphs 2020 Shared Task: Leveraging Pretrained Language Models for Explanation Regeneration 共享任务:利用预训练语言模型进行解释再生
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.12
Aditya Girish Pawate, Varun Madhavan, Devansh Chandak
{"title":"ChiSquareX at TextGraphs 2020 Shared Task: Leveraging Pretrained Language Models for Explanation Regeneration","authors":"Aditya Girish Pawate, Varun Madhavan, Devansh Chandak","doi":"10.18653/v1/2020.textgraphs-1.12","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.12","url":null,"abstract":"In this work, we describe the system developed by a group of undergraduates from the Indian Institutes of Technology for the Shared Task at TextGraphs-14 on Multi-Hop Inference Explanation Regeneration (Jansen and Ustalov, 2020). The shared task required participants to develop methods to reconstruct gold explanations for elementary science questions from the WorldTreeCorpus (Xie et al., 2020). Although our research was not funded by any organization and all the models were trained on freely available tools like Google Colab, which restricted our computational capabilities, we have managed to achieve noteworthy results, placing ourselves in 4th place with a MAPscore of 0.49021in the evaluation leaderboard and 0.5062 MAPscore on the post-evaluation-phase leaderboard using RoBERTa. We incorporated some of the methods proposed in the previous edition of Textgraphs-13 (Chia et al., 2019), which proved to be very effective, improved upon them, and built a model on top of it using powerful state-of-the-art pre-trained language models like RoBERTa (Liu et al., 2019), BART (Lewis et al., 2020), SciB-ERT (Beltagy et al., 2019) among others. Further optimization of our work can be done with the availability of better computational resources.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123172118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A survey of embedding models of entities and relationships for knowledge graph completion 知识图谱补全中实体和关系的嵌入模型综述
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.1
Dat Quoc Nguyen
{"title":"A survey of embedding models of entities and relationships for knowledge graph completion","authors":"Dat Quoc Nguyen","doi":"10.18653/v1/2020.textgraphs-1.1","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.1","url":null,"abstract":"Knowledge graphs (KGs) of real-world facts about entities and their relationships are useful resources for a variety of natural language processing tasks. However, because knowledge graphs are typically incomplete, it is useful to perform knowledge graph completion or link prediction, i.e. predict whether a relationship not in the knowledge graph is likely to be true. This paper serves as a comprehensive survey of embedding models of entities and relationships for knowledge graph completion, summarizing up-to-date experimental results on standard benchmark datasets and pointing out potential future research directions.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114220531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
PGL at TextGraphs 2020 Shared Task: Explanation Regeneration using Language and Graph Learning Methods PGL在TextGraphs 2020共享任务:使用语言和图形学习方法的解释再生
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.11
Weibin Li, Yuxiang Lu, Zhengjie Huang, Weiyue Su, Jiaxiang Liu, Shikun Feng, Yu Sun
{"title":"PGL at TextGraphs 2020 Shared Task: Explanation Regeneration using Language and Graph Learning Methods","authors":"Weibin Li, Yuxiang Lu, Zhengjie Huang, Weiyue Su, Jiaxiang Liu, Shikun Feng, Yu Sun","doi":"10.18653/v1/2020.textgraphs-1.11","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.11","url":null,"abstract":"This paper describes the system designed by the Baidu PGL Team which achieved the first place in the TextGraphs 2020 Shared Task. The task focuses on generating explanations for elementary science questions. Given a question and its corresponding correct answer, we are asked to select the facts that can explain why the answer is correct for the question and answering (QA) from a large knowledge base. To address this problem, we use a pre-trained language model to recall the top-K relevant explanations for each question. Then, we adopt a re-ranking approach based on a pre-trained language model to rank the candidate explanations. To further improve the rankings, we also develop an architecture consisting both powerful pre-trained transformers and GNNs to tackle the multi-hop inference problem. The official evaluation shows that, our system can outperform the second best system by 1.91 points.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132706132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Graph-based Syntactic Word Embeddings 基于图的句法词嵌入
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.8
Ragheb Al-Ghezi, M. Kurimo
{"title":"Graph-based Syntactic Word Embeddings","authors":"Ragheb Al-Ghezi, M. Kurimo","doi":"10.18653/v1/2020.textgraphs-1.8","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.8","url":null,"abstract":"We propose a simple and efficient framework to learn syntactic embeddings based on information derived from constituency parse trees. Using biased random walk methods, our embeddings not only encode syntactic information about words, but they also capture contextual information. We also propose a method to train the embeddings on multiple constituency parse trees to ensure the encoding of global syntactic representation. Quantitative evaluation of the embeddings show a competitive performance on POS tagging task when compared to other types of embeddings, and qualitative evaluation reveals interesting facts about the syntactic typology learned by these embeddings.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133482313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Relation Specific Transformations for Open World Knowledge Graph Completion 面向开放世界知识图谱补全的关系特定转换
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.9
Haseeb Shah, Johannes Villmow, A. Ulges
{"title":"Relation Specific Transformations for Open World Knowledge Graph Completion","authors":"Haseeb Shah, Johannes Villmow, A. Ulges","doi":"10.18653/v1/2020.textgraphs-1.9","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.9","url":null,"abstract":"We propose an open-world knowledge graph completion model that can be combined with common closed-world approaches (such as ComplEx) and enhance them to exploit text-based representations for entities unseen in training. Our model learns relation-specific transformation functions from text-based to graph-based embedding space, where the closed-world link prediction model can be applied. We demonstrate state-of-the-art results on common open-world benchmarks and show that our approach benefits from relation-specific transformation functions (RST), giving substantial improvements over a relation-agnostic approach.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131242661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Merge and Recognize: A Geometry and 2D Context Aware Graph Model for Named Entity Recognition from Visual Documents 合并与识别:一种用于可视化文档命名实体识别的几何和2D上下文感知图形模型
Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs) Pub Date : 1900-01-01 DOI: 10.18653/v1/2020.textgraphs-1.3
Chuwei Luo, Yongpan Wang, Qi Zheng, Liangcheng Li, Feiyu Gao, Shiyu Zhang
{"title":"Merge and Recognize: A Geometry and 2D Context Aware Graph Model for Named Entity Recognition from Visual Documents","authors":"Chuwei Luo, Yongpan Wang, Qi Zheng, Liangcheng Li, Feiyu Gao, Shiyu Zhang","doi":"10.18653/v1/2020.textgraphs-1.3","DOIUrl":"https://doi.org/10.18653/v1/2020.textgraphs-1.3","url":null,"abstract":"Named entity recognition (NER) from visual documents, such as invoices, receipts or business cards, is a critical task for visual document understanding. Most classical approaches use a sequence-based model (typically BiLSTM-CRF framework) without considering document structure. Recent work on graph-based model using graph convolutional networks to encode visual and textual features have achieved promising performance on the task. However, few attempts take geometry information of text segments (text in bounding box) in visual documents into account. Meanwhile, existing methods do not consider that related text segments which need to be merged to form a complete entity in many real-world situations. In this paper, we present GraphNEMR, a graph-based model that uses graph convolutional networks to jointly merge text segments and recognize named entities. By incorporating geometry information from visual documents into our model, richer 2D context information is generated to improve document representations. To merge text segments, we introduce a novel mechanism that captures both geometry information as well as semantic information based on pre-trained language model. Experimental results show that the proposed GraphNEMR model outperforms both sequence-based and graph-based SOTA methods significantly.","PeriodicalId":282839,"journal":{"name":"Proceedings of the Graph-based Methods for Natural Language Processing (TextGraphs)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122893222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信