{"title":"Alignment Offset Based Adaptive Training for Simultaneous Machine Translation","authors":"Qiqi Liang, Yanjun Liu, Fandong Meng, Jinan Xu, Yufeng Chen, Jie Zhou","doi":"10.1109/ICNLP58431.2023.00035","DOIUrl":null,"url":null,"abstract":"Given incomplete source sentences as inputs, it is generally difficult for Simultaneous Machine Translation (SiMT) models to generate a target token once its aligned source tokens are absent. How to measure such difficulty and further conduct adaptive training for SiMT models are not sufficiently studied. In this paper, we propose a new metric named alignment offset (AO) to quantify the learning difficulty of target tokens for SiMT models. Given a target token, its AO is calculated by the offset between its aligned source tokens and the already received source tokens. Furthermore, we design two AO-based adaptive training methods to improve the training of SiMT models. Firstly, we introduce token-level curriculum learning based on AO, which progressively switches the training process from easy target tokens to difficult ones. Secondly, we assign an appropriate weight to the training loss of each target token according to its AO. Experimental results on four datasets demonstrate that our methods significantly and consistently outperform all the strong baselines.","PeriodicalId":53637,"journal":{"name":"Icon","volume":"28 1","pages":"157-164"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNLP58431.2023.00035","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
引用次数: 0
Abstract
Given incomplete source sentences as inputs, it is generally difficult for Simultaneous Machine Translation (SiMT) models to generate a target token once its aligned source tokens are absent. How to measure such difficulty and further conduct adaptive training for SiMT models are not sufficiently studied. In this paper, we propose a new metric named alignment offset (AO) to quantify the learning difficulty of target tokens for SiMT models. Given a target token, its AO is calculated by the offset between its aligned source tokens and the already received source tokens. Furthermore, we design two AO-based adaptive training methods to improve the training of SiMT models. Firstly, we introduce token-level curriculum learning based on AO, which progressively switches the training process from easy target tokens to difficult ones. Secondly, we assign an appropriate weight to the training loss of each target token according to its AO. Experimental results on four datasets demonstrate that our methods significantly and consistently outperform all the strong baselines.