Automatic Numerical Question Answering on Table using BERT-GNN

Ruchi Bagwe, K. George
{"title":"Automatic Numerical Question Answering on Table using BERT-GNN","authors":"Ruchi Bagwe, K. George","doi":"10.1109/UEMCON51285.2020.9298028","DOIUrl":null,"url":null,"abstract":"The table base numerical question-answering task requires a mechanism to understand the relation between table content and numbers (present in table and question). It also needs an efficient method to address complex reasoning on table context. Most of the existing approaches in the natural language processing technology address the context-based questions on the table but fail to address the numerical reasoning part. They are also built on a large search database, which makes it challenging to use them in multiple domains. These approaches use pre-trained models like BERT to perform context encoding of a complete table. Hence these models fail when a large table is provided as input, as full table encoding is a very resource and time-consuming task. In this paper, a framework is proposed to answer questions on the table with numerical reasoning. This framework uses a context-snapshot mechanism to filter irrelevant table rows before tokenizing the table content. The filtered context and tokenized question are converted into vector representation using a pre-trained BERT model. This proposed model finds the correlation between the tokenized context-snapshot and numbers in question using graph neural networks. Further, it uses a feed-forward neural network to perform the numerical operation to compute the answer. The model is trained and evaluated on WikiTableQuestions datasets, shows a promising result.","PeriodicalId":433609,"journal":{"name":"2020 11th IEEE Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 11th IEEE Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/UEMCON51285.2020.9298028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

The table base numerical question-answering task requires a mechanism to understand the relation between table content and numbers (present in table and question). It also needs an efficient method to address complex reasoning on table context. Most of the existing approaches in the natural language processing technology address the context-based questions on the table but fail to address the numerical reasoning part. They are also built on a large search database, which makes it challenging to use them in multiple domains. These approaches use pre-trained models like BERT to perform context encoding of a complete table. Hence these models fail when a large table is provided as input, as full table encoding is a very resource and time-consuming task. In this paper, a framework is proposed to answer questions on the table with numerical reasoning. This framework uses a context-snapshot mechanism to filter irrelevant table rows before tokenizing the table content. The filtered context and tokenized question are converted into vector representation using a pre-trained BERT model. This proposed model finds the correlation between the tokenized context-snapshot and numbers in question using graph neural networks. Further, it uses a feed-forward neural network to perform the numerical operation to compute the answer. The model is trained and evaluated on WikiTableQuestions datasets, shows a promising result.
基于BERT-GNN的表格自动数值答题
基于表的数字问答任务需要一种机制来理解表内容和数字(出现在表和问题中)之间的关系。它还需要一种有效的方法来处理表上下文的复杂推理。在自然语言处理技术中,现有的大多数方法都解决了基于上下文的表格问题,但未能解决数字推理部分。它们还建立在一个大型搜索数据库上,这使得在多个领域使用它们具有挑战性。这些方法使用BERT等预训练模型来执行完整表的上下文编码。因此,当提供一个大表作为输入时,这些模型会失败,因为全表编码是一项非常耗费资源和时间的任务。本文提出了一个用数值推理来回答表格上的问题的框架。该框架使用上下文快照机制在对表内容进行标记之前过滤不相关的表行。过滤后的上下文和标记化的问题使用预训练的BERT模型转换为向量表示。该模型使用图神经网络发现标记化的上下文快照与有问题的数字之间的相关性。此外,它使用前馈神经网络执行数值运算来计算答案。该模型在WikiTableQuestions数据集上进行了训练和评估,显示出良好的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信