{"title":"Mizo视觉基因组1.0:英语-Mizo多模态神经机器翻译数据集","authors":"Vanlalmuansangi Khenglawt, Sahinur Rahman Laskar, Riyanka Manna, Partha Pakray, Ajoy Kumar Khan","doi":"10.1109/SILCON55242.2022.10028882","DOIUrl":null,"url":null,"abstract":"Multimodal machine translation (MMT) handles extracting information from several modalities, considering the presumption that the extra modalities will include beneficial alternative perspectives of the input data. Regardless of its significant benefits, it is challenging to implement an MMT system for several languages, mainly due to the scarcity of the availability of multimodal datasets. As for the low-resource English-Mizo pair, the standard multimodal corpus is not available. Therefore, in this paper, we have developed a Mizo Visual Genome 1.0 (MVG 1.0) dataset for English-Mizo MMT, including images with corresponding bilingual textual descriptions. According to automated assessment measures, the performance of multimodal neural machine translation (MNMT) is better than text-only neural machine translation. To the best of our knowledge, our English-Mizo MMT system is the pioneering work in this approach, and as such, it can serve as a baseline for future study in MMT for the low-resource English-Mizo language pair.","PeriodicalId":183947,"journal":{"name":"2022 IEEE Silchar Subsection Conference (SILCON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Mizo Visual Genome 1.0 : A Dataset for English-Mizo Multimodal Neural Machine Translation\",\"authors\":\"Vanlalmuansangi Khenglawt, Sahinur Rahman Laskar, Riyanka Manna, Partha Pakray, Ajoy Kumar Khan\",\"doi\":\"10.1109/SILCON55242.2022.10028882\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multimodal machine translation (MMT) handles extracting information from several modalities, considering the presumption that the extra modalities will include beneficial alternative perspectives of the input data. Regardless of its significant benefits, it is challenging to implement an MMT system for several languages, mainly due to the scarcity of the availability of multimodal datasets. As for the low-resource English-Mizo pair, the standard multimodal corpus is not available. Therefore, in this paper, we have developed a Mizo Visual Genome 1.0 (MVG 1.0) dataset for English-Mizo MMT, including images with corresponding bilingual textual descriptions. According to automated assessment measures, the performance of multimodal neural machine translation (MNMT) is better than text-only neural machine translation. To the best of our knowledge, our English-Mizo MMT system is the pioneering work in this approach, and as such, it can serve as a baseline for future study in MMT for the low-resource English-Mizo language pair.\",\"PeriodicalId\":183947,\"journal\":{\"name\":\"2022 IEEE Silchar Subsection Conference (SILCON)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE Silchar Subsection Conference (SILCON)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SILCON55242.2022.10028882\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE Silchar Subsection Conference (SILCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SILCON55242.2022.10028882","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Mizo Visual Genome 1.0 : A Dataset for English-Mizo Multimodal Neural Machine Translation
Multimodal machine translation (MMT) handles extracting information from several modalities, considering the presumption that the extra modalities will include beneficial alternative perspectives of the input data. Regardless of its significant benefits, it is challenging to implement an MMT system for several languages, mainly due to the scarcity of the availability of multimodal datasets. As for the low-resource English-Mizo pair, the standard multimodal corpus is not available. Therefore, in this paper, we have developed a Mizo Visual Genome 1.0 (MVG 1.0) dataset for English-Mizo MMT, including images with corresponding bilingual textual descriptions. According to automated assessment measures, the performance of multimodal neural machine translation (MNMT) is better than text-only neural machine translation. To the best of our knowledge, our English-Mizo MMT system is the pioneering work in this approach, and as such, it can serve as a baseline for future study in MMT for the low-resource English-Mizo language pair.