Vietnamese Question Answering System f rom Multilingual BERT Models to Monolingual BERT Model

Nguyen Thi Mai Trang, M. Shcherbakov
{"title":"Vietnamese Question Answering System f rom Multilingual BERT Models to Monolingual BERT Model","authors":"Nguyen Thi Mai Trang, M. Shcherbakov","doi":"10.1109/SMART50582.2020.9337155","DOIUrl":null,"url":null,"abstract":"A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.","PeriodicalId":129946,"journal":{"name":"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SMART50582.2020.9337155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.
从多语言BERT模型到单语言BERT模型的越南语问答系统
基于自然语言处理和深度学习的问答系统越来越受到人工智能社区的关注。许多公司和组织都对开发自动问答系统感兴趣,这一系统正在得到广泛的研究。近年来,为了解决自然语言处理任务的局限性,提出了一种新的模型——双向编码器转换表示(BERT)。BERT在几乎所有包括QA任务的任务中都取得了最好的结果。在这项工作中,我们尝试应用多语言BERT模型(多语言BERT [1], DeepPavlov多语言BERT, XQuAD上微调的多语言BERT)和越南语特定语言的BERT模型(PhoBERT)。结果表明,单语言模型优于多语言模型。我们还建议在XQuAD模型上进行多语言BERT微调,如果系统是基于多语言BERT模型构建的,则作为构建越南QA系统的选项。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信