{"title":"DA-BERT: Enhancing Knowledge Selection in Dialog via Domain Adapted BERT with Dynamic Masking Probability","authors":"Zhiguo Zeng, Chi-Yin Chow, Ning Li","doi":"10.1109/SMARTCOMP52413.2021.00040","DOIUrl":null,"url":null,"abstract":"One of the most challenging tasks in Knowledge-grounded Task-oriented Dialog Systems (KTDS) is the knowledge selection task, which aims to find the proper knowledge snippets to handle user requests. This paper proposes DA-BERT to employ pre-trained BERT with domain adaptive training and newly proposed dynamic masking probability to deal with knowledge selection in KTDS. Domain adaptive training minimizes the domain gap between the general text data BERT is pre-trained on and the dialog-knowledge joint data; and dynamic masking probability enhances the training in an easy-to-hard mode. Experimental results on the benchmark dataset show that our proposed training method outperforms the state-of-the-art models with large margins across all the evaluation metrics. Moreover, we analyze the bad case of our method and recognize several typical errors in the bad case set to facilitate further research in this direction.","PeriodicalId":330785,"journal":{"name":"2021 IEEE International Conference on Smart Computing (SMARTCOMP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Smart Computing (SMARTCOMP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SMARTCOMP52413.2021.00040","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
One of the most challenging tasks in Knowledge-grounded Task-oriented Dialog Systems (KTDS) is the knowledge selection task, which aims to find the proper knowledge snippets to handle user requests. This paper proposes DA-BERT to employ pre-trained BERT with domain adaptive training and newly proposed dynamic masking probability to deal with knowledge selection in KTDS. Domain adaptive training minimizes the domain gap between the general text data BERT is pre-trained on and the dialog-knowledge joint data; and dynamic masking probability enhances the training in an easy-to-hard mode. Experimental results on the benchmark dataset show that our proposed training method outperforms the state-of-the-art models with large margins across all the evaluation metrics. Moreover, we analyze the bad case of our method and recognize several typical errors in the bad case set to facilitate further research in this direction.