Open-domain Conversational Agent based on Pre-trained Transformers for Human-Robot Interaction

M. Fernandes, Plinio Moreno
{"title":"Open-domain Conversational Agent based on Pre-trained Transformers for Human-Robot Interaction","authors":"M. Fernandes, Plinio Moreno","doi":"10.5220/0011300800003277","DOIUrl":null,"url":null,"abstract":": Generative pre-trained transformers belong to the breakthroughs in Natural Language Processing (NLP), allowing Human-Robot Interactions ( e.g. the creation of an open-domain chatbot). However, a substantial amount of research and available data are in English, causing low-resourced languages to be overlooked. This work addresses this problem for European Portuguese with two options: (i) Translation of the sentences before and after using the model fine-tuned on an English-based dataset, (ii) Translation of the English-based dataset to Portuguese and then fine-tune this model on it. We rely on the DialoGPT (dialogue generative pre-trained transformer), a tunable neural conversational answer generation model that learns the basic skills to conduct a dialogue. We use two sources of evaluation: (i) Metrics for text generation based on uncertainty ( i.e. perplexity), and similarity between sentences ( i.e. BLEU, METEOR and ROUGE) and (ii) Human-based evaluation of the sentences. The translation of sentences before and after of the modified DialoGPT model, using the Daily Dialogue dataset led to the best results.","PeriodicalId":88612,"journal":{"name":"News. Phi Delta Epsilon","volume":"1 1","pages":"168-175"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"News. Phi Delta Epsilon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5220/0011300800003277","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

: Generative pre-trained transformers belong to the breakthroughs in Natural Language Processing (NLP), allowing Human-Robot Interactions ( e.g. the creation of an open-domain chatbot). However, a substantial amount of research and available data are in English, causing low-resourced languages to be overlooked. This work addresses this problem for European Portuguese with two options: (i) Translation of the sentences before and after using the model fine-tuned on an English-based dataset, (ii) Translation of the English-based dataset to Portuguese and then fine-tune this model on it. We rely on the DialoGPT (dialogue generative pre-trained transformer), a tunable neural conversational answer generation model that learns the basic skills to conduct a dialogue. We use two sources of evaluation: (i) Metrics for text generation based on uncertainty ( i.e. perplexity), and similarity between sentences ( i.e. BLEU, METEOR and ROUGE) and (ii) Human-based evaluation of the sentences. The translation of sentences before and after of the modified DialoGPT model, using the Daily Dialogue dataset led to the best results.
基于预训练变压器的人机交互开放域会话代理
生成式预训练转换器属于自然语言处理(NLP)的突破,允许人机交互(例如创建开放域聊天机器人)。然而,大量的研究和可用数据都是英文的,导致资源匮乏的语言被忽视。这项工作通过两种选择解决了欧洲葡萄牙语的这个问题:(i)在基于英语的数据集上使用模型微调之前和之后翻译句子,(ii)将基于英语的数据集翻译成葡萄牙语,然后在其上微调该模型。我们依靠DialoGPT(对话生成预训练转换器),这是一个可调的神经会话答案生成模型,它学习进行对话的基本技能。我们使用两种评估来源:(i)基于不确定性(即困惑)和句子之间的相似性(即BLEU, METEOR和ROUGE)的文本生成度量;(ii)基于人的句子评估。使用Daily Dialogue数据集对改进的DialoGPT模型前后的句子进行翻译,结果最好。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信