{"title":"Vietnamese Question Answering System f rom Multilingual BERT Models to Monolingual BERT Model","authors":"Nguyen Thi Mai Trang, M. Shcherbakov","doi":"10.1109/SMART50582.2020.9337155","DOIUrl":null,"url":null,"abstract":"A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.","PeriodicalId":129946,"journal":{"name":"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SMART50582.2020.9337155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.