{"title":"芬兰语-瑞典语翻译的bert融合模型","authors":"I. Kumpulainen, J. Vankka","doi":"10.1109/SNAMS53716.2021.9731849","DOIUrl":null,"url":null,"abstract":"Translation between Finnish and Swedish is a common yet time-consuming and expensive task. In this paper, we train new neural machine translation models and compare them with publicly available tools for automatic translation of Finnish to Swedish. Furthermore, we analyze if fusing BERT models with traditional Transformer models produces better translations. We train a base Transformer and a large Transformer model using Fairseq and compare the results with BERT-fused versions of the models. Our large transformer model matches the state-of-the-art performance in Finnish-Swedish translation and slightly improves the BLEU score from 29.4 to 29.8. In our experiments, fusing the smaller Transformer model with a pre-trained BERT improves the quality of the translations. Surprisingly, the larger Transformer model in contrast does not benefit from being fused with a BERT model.","PeriodicalId":387260,"journal":{"name":"2021 Eighth International Conference on Social Network Analysis, Management and Security (SNAMS)","volume":"112 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"BERT-fused Model for Finnish-Swedish Translation\",\"authors\":\"I. Kumpulainen, J. Vankka\",\"doi\":\"10.1109/SNAMS53716.2021.9731849\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Translation between Finnish and Swedish is a common yet time-consuming and expensive task. In this paper, we train new neural machine translation models and compare them with publicly available tools for automatic translation of Finnish to Swedish. Furthermore, we analyze if fusing BERT models with traditional Transformer models produces better translations. We train a base Transformer and a large Transformer model using Fairseq and compare the results with BERT-fused versions of the models. Our large transformer model matches the state-of-the-art performance in Finnish-Swedish translation and slightly improves the BLEU score from 29.4 to 29.8. In our experiments, fusing the smaller Transformer model with a pre-trained BERT improves the quality of the translations. Surprisingly, the larger Transformer model in contrast does not benefit from being fused with a BERT model.\",\"PeriodicalId\":387260,\"journal\":{\"name\":\"2021 Eighth International Conference on Social Network Analysis, Management and Security (SNAMS)\",\"volume\":\"112 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 Eighth International Conference on Social Network Analysis, Management and Security (SNAMS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SNAMS53716.2021.9731849\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 Eighth International Conference on Social Network Analysis, Management and Security (SNAMS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SNAMS53716.2021.9731849","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Translation between Finnish and Swedish is a common yet time-consuming and expensive task. In this paper, we train new neural machine translation models and compare them with publicly available tools for automatic translation of Finnish to Swedish. Furthermore, we analyze if fusing BERT models with traditional Transformer models produces better translations. We train a base Transformer and a large Transformer model using Fairseq and compare the results with BERT-fused versions of the models. Our large transformer model matches the state-of-the-art performance in Finnish-Swedish translation and slightly improves the BLEU score from 29.4 to 29.8. In our experiments, fusing the smaller Transformer model with a pre-trained BERT improves the quality of the translations. Surprisingly, the larger Transformer model in contrast does not benefit from being fused with a BERT model.