Learning to Correct Erroneous Words for Document Grounded Conversations

Junyan Qiu, Haidong Zhang, Yiping Yang
{"title":"Learning to Correct Erroneous Words for Document Grounded Conversations","authors":"Junyan Qiu, Haidong Zhang, Yiping Yang","doi":"10.1145/3587828.3587883","DOIUrl":null,"url":null,"abstract":"Document grounded conversation (DGC) aims to generate informative responses when talking about a document. It is normally formulated as a sequence-to-sequence (Seq2seq) learning problem, which directly maps source sequences, i.e., the context and background documents, to the target sequence, i.e., the response. These responses are normally used as the final output without further polishing, which may suffer from the global information loss owing to the auto-regression paradigm. To tackle this problem, some researches designed two-pass generation to improve the quality of responses. However, these approaches lack the capability of distinguishing inappropriate words in the first pass, which may maintain the erroneous words while rewrite the correct ones. In this paper, we design a scheduled error correction network (SECN) with multiple generation passes to explicitly locate and rewrite the erroneous words in previous passes. Specifically, a discriminator is employed to distinguish erroneous words which are further revised by a refiner. Moreover, we also apply curriculum learning with reasonable learning schedule to train our model from easy to hard conversations, where the complexity is measured by the number of decoding passes. We conduct comprehensive experiments on a public document grounded conversation dataset, Wizard-of-Wikipedia, and the results demonstrate significant promotions over several strong benchmarks.","PeriodicalId":340917,"journal":{"name":"Proceedings of the 2023 12th International Conference on Software and Computer Applications","volume":"221 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 12th International Conference on Software and Computer Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3587828.3587883","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Document grounded conversation (DGC) aims to generate informative responses when talking about a document. It is normally formulated as a sequence-to-sequence (Seq2seq) learning problem, which directly maps source sequences, i.e., the context and background documents, to the target sequence, i.e., the response. These responses are normally used as the final output without further polishing, which may suffer from the global information loss owing to the auto-regression paradigm. To tackle this problem, some researches designed two-pass generation to improve the quality of responses. However, these approaches lack the capability of distinguishing inappropriate words in the first pass, which may maintain the erroneous words while rewrite the correct ones. In this paper, we design a scheduled error correction network (SECN) with multiple generation passes to explicitly locate and rewrite the erroneous words in previous passes. Specifically, a discriminator is employed to distinguish erroneous words which are further revised by a refiner. Moreover, we also apply curriculum learning with reasonable learning schedule to train our model from easy to hard conversations, where the complexity is measured by the number of decoding passes. We conduct comprehensive experiments on a public document grounded conversation dataset, Wizard-of-Wikipedia, and the results demonstrate significant promotions over several strong benchmarks.
学习纠正基于文件的对话中的错误单词
基于文档的对话(DGC)的目的是在谈论文档时生成信息丰富的响应。它通常被表述为序列到序列(Seq2seq)学习问题,它直接将源序列(即上下文和背景文档)映射到目标序列(即响应)。这些响应通常用作最终输出,而不进行进一步的修饰,这可能会由于自回归范式而遭受全局信息损失。为了解决这个问题,一些研究设计了两遍生成来提高响应的质量。然而,这些方法缺乏第一次识别不恰当词的能力,这可能会保留错误的词,而重写正确的词。在本文中,我们设计了一个具有多个生成通道的定时纠错网络(SECN),以显式定位和重写先前通道中的错误单词。具体地说,使用一个鉴别器来区分错别字,这些错别字由精炼器进一步修正。此外,我们还应用课程学习和合理的学习计划来训练我们的模型,从简单到困难的对话,其中复杂性是通过解码通过的次数来衡量的。我们在基于对话数据集Wizard-of-Wikipedia的公共文档上进行了全面的实验,结果表明在几个强大的基准测试中有显著的提升。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信