Learning to Select the Relevant History Turns in Conversational Question Answering

WISE Pub Date : 2023-08-04 DOI:10.48550/arXiv.2308.02294
Munazza Zaib, Wei Emma Zhang, Quan Z. Sheng, S. Sagar, A. Mahmood, Yang Zhang
{"title":"Learning to Select the Relevant History Turns in Conversational Question Answering","authors":"Munazza Zaib, Wei Emma Zhang, Quan Z. Sheng, S. Sagar, A. Mahmood, Yang Zhang","doi":"10.48550/arXiv.2308.02294","DOIUrl":null,"url":null,"abstract":"The increasing demand for the web-based digital assistants has given a rapid rise in the interest of the Information Retrieval (IR) community towards the field of conversational question answering (ConvQA). However, one of the critical aspects of ConvQA is the effective selection of conversational history turns to answer the question at hand. The dependency between relevant history selection and correct answer prediction is an intriguing but under-explored area. The selected relevant context can better guide the system so as to where exactly in the passage to look for an answer. Irrelevant context, on the other hand, brings noise to the system, thereby resulting in a decline in the model's performance. In this paper, we propose a framework, DHS-ConvQA (Dynamic History Selection in Conversational Question Answering), that first generates the context and question entities for all the history turns, which are then pruned on the basis of similarity they share in common with the question at hand. We also propose an attention-based mechanism to re-rank the pruned terms based on their calculated weights of how useful they are in answering the question. In the end, we further aid the model by highlighting the terms in the re-ranked conversational history using a binary classification task and keeping the useful terms (predicted as 1) and ignoring the irrelevant terms (predicted as 0). We demonstrate the efficacy of our proposed framework with extensive experimental results on CANARD and QuAC -- the two popularly utilized datasets in ConvQA. We demonstrate that selecting relevant turns works better than rewriting the original question. We also investigate how adding the irrelevant history turns negatively impacts the model's performance and discuss the research challenges that demand more attention from the IR community.","PeriodicalId":424892,"journal":{"name":"WISE","volume":"89 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"WISE","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2308.02294","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The increasing demand for the web-based digital assistants has given a rapid rise in the interest of the Information Retrieval (IR) community towards the field of conversational question answering (ConvQA). However, one of the critical aspects of ConvQA is the effective selection of conversational history turns to answer the question at hand. The dependency between relevant history selection and correct answer prediction is an intriguing but under-explored area. The selected relevant context can better guide the system so as to where exactly in the passage to look for an answer. Irrelevant context, on the other hand, brings noise to the system, thereby resulting in a decline in the model's performance. In this paper, we propose a framework, DHS-ConvQA (Dynamic History Selection in Conversational Question Answering), that first generates the context and question entities for all the history turns, which are then pruned on the basis of similarity they share in common with the question at hand. We also propose an attention-based mechanism to re-rank the pruned terms based on their calculated weights of how useful they are in answering the question. In the end, we further aid the model by highlighting the terms in the re-ranked conversational history using a binary classification task and keeping the useful terms (predicted as 1) and ignoring the irrelevant terms (predicted as 0). We demonstrate the efficacy of our proposed framework with extensive experimental results on CANARD and QuAC -- the two popularly utilized datasets in ConvQA. We demonstrate that selecting relevant turns works better than rewriting the original question. We also investigate how adding the irrelevant history turns negatively impacts the model's performance and discuss the research challenges that demand more attention from the IR community.
学习在会话式问答中选择相关的历史转折
对基于网络的数字助理的需求日益增长,使得信息检索(IR)社区对会话问答(ConvQA)领域的兴趣迅速上升。然而,ConvQA的一个关键方面是有效地选择会话历史来回答手头的问题。相关历史选择和正确答案预测之间的依赖关系是一个有趣但尚未充分探索的领域。所选择的相关上下文可以更好地引导系统,以便准确地在文章中寻找答案。另一方面,不相关的上下文会给系统带来噪声,从而导致模型的性能下降。在本文中,我们提出了一个框架,DHS-ConvQA(会话问答中的动态历史选择),它首先生成所有历史回合的上下文和问题实体,然后根据它们与手头问题共有的相似性对它们进行修剪。我们还提出了一种基于注意力的机制,根据它们在回答问题时的有用程度的计算权重来重新排列修剪后的术语。最后,我们通过使用二元分类任务突出显示重新排序的会话历史中的术语,并保留有用的术语(预测为1)并忽略无关的术语(预测为0)来进一步帮助模型。我们通过在CANARD和QuAC (ConvQA中常用的两个数据集)上的大量实验结果证明了我们提出的框架的有效性。我们证明了选择相关的回合比重写原始问题效果更好。我们还研究了添加不相关历史转折如何对模型的性能产生负面影响,并讨论了需要IR社区更多关注的研究挑战。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信