{"title":"通过多任务提高DST模型的少镜头性能,更好地为语言障碍人士服务","authors":"Mingyang Sun, QiXiang Gao, Yutao Mou, Guanting Dong, Ruifang Liu, Wenbin Guo","doi":"10.1109/ICASSPW59220.2023.10193387","DOIUrl":null,"url":null,"abstract":"Artificial intelligence-based virtual assistants make people’s daily life more convenient. However, the utterances of language-impaired people are limited and different in characteristics and domains from that of ordinary people. So it is difficult for language-impaired people to benefit from standard data-driven artificial intelligence algorithms. In this paper, we propose a multi-task training method for the dialogue state tracking (DST) task in dialogue systems that make up virtual assistants, improving the performance of T5 on the few-shot cross-domain DST task. Specifically, we consider two ways of handling DST task: predicting the dialogue state from the beginning or updating the dialogue state every turn, and accordingly design the main task and auxiliary task for the model. Experiments show that our method outperforms most previous works on the MultiWOZ 2.0 and 2.1 datasets for the few-shot cross-domain DST task. For the artificial-crafted language-impaired dataset, our method can effectively improve the few-shot cross-domain performance of the model. Additionally, we analyzed the possible reason why this multitasking approach works well.","PeriodicalId":158726,"journal":{"name":"2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Improving Few-Shot Performance of DST Model Through Multitask to Better Serve Language-Impaired People\",\"authors\":\"Mingyang Sun, QiXiang Gao, Yutao Mou, Guanting Dong, Ruifang Liu, Wenbin Guo\",\"doi\":\"10.1109/ICASSPW59220.2023.10193387\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Artificial intelligence-based virtual assistants make people’s daily life more convenient. However, the utterances of language-impaired people are limited and different in characteristics and domains from that of ordinary people. So it is difficult for language-impaired people to benefit from standard data-driven artificial intelligence algorithms. In this paper, we propose a multi-task training method for the dialogue state tracking (DST) task in dialogue systems that make up virtual assistants, improving the performance of T5 on the few-shot cross-domain DST task. Specifically, we consider two ways of handling DST task: predicting the dialogue state from the beginning or updating the dialogue state every turn, and accordingly design the main task and auxiliary task for the model. Experiments show that our method outperforms most previous works on the MultiWOZ 2.0 and 2.1 datasets for the few-shot cross-domain DST task. For the artificial-crafted language-impaired dataset, our method can effectively improve the few-shot cross-domain performance of the model. Additionally, we analyzed the possible reason why this multitasking approach works well.\",\"PeriodicalId\":158726,\"journal\":{\"name\":\"2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)\",\"volume\":\"65 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICASSPW59220.2023.10193387\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICASSPW59220.2023.10193387","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Improving Few-Shot Performance of DST Model Through Multitask to Better Serve Language-Impaired People
Artificial intelligence-based virtual assistants make people’s daily life more convenient. However, the utterances of language-impaired people are limited and different in characteristics and domains from that of ordinary people. So it is difficult for language-impaired people to benefit from standard data-driven artificial intelligence algorithms. In this paper, we propose a multi-task training method for the dialogue state tracking (DST) task in dialogue systems that make up virtual assistants, improving the performance of T5 on the few-shot cross-domain DST task. Specifically, we consider two ways of handling DST task: predicting the dialogue state from the beginning or updating the dialogue state every turn, and accordingly design the main task and auxiliary task for the model. Experiments show that our method outperforms most previous works on the MultiWOZ 2.0 and 2.1 datasets for the few-shot cross-domain DST task. For the artificial-crafted language-impaired dataset, our method can effectively improve the few-shot cross-domain performance of the model. Additionally, we analyzed the possible reason why this multitasking approach works well.