Proceedings of the Third Workshop on Automatic Simultaneous Translation最新文献

筛选
英文 中文
USST’s System for AutoSimTrans 2022 USST的AutoSimTrans系统2022
Proceedings of the Third Workshop on Automatic Simultaneous Translation Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.autosimtrans-1.7
Zhu Hui, Yu Jun
{"title":"USST’s System for AutoSimTrans 2022","authors":"Zhu Hui, Yu Jun","doi":"10.18653/v1/2022.autosimtrans-1.7","DOIUrl":"https://doi.org/10.18653/v1/2022.autosimtrans-1.7","url":null,"abstract":"This paper describes our submitted text-to-text Simultaneous translation (ST) system, which won the second place in the Chinese→English streaming translation task of AutoSimTrans 2022. Our baseline system is a BPE-based Transformer model trained with the PaddlePaddle framework. In our experiments, we employ data synthesis and ensemble approaches to enhance the base model. In order to bridge the gap between general domain and spoken domain, we select in-domain data from general corpus and mixed then with spoken corpus for mixed fine tuning. Finally, we adopt fixed wait-k policy to transfer our full-sentence translation model to simultaneous translation model. Experiments on the development data show that our system outperforms than the baseline system.","PeriodicalId":444422,"journal":{"name":"Proceedings of the Third Workshop on Automatic Simultaneous Translation","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133385495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Findings of the Third Workshop on Automatic Simultaneous Translation 第三届自动同声翻译研讨会的研究结果
Proceedings of the Third Workshop on Automatic Simultaneous Translation Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.autosimtrans-1.1
Ruiqing Zhang, Chuanqiang Zhang, Zhongjun He, Hua Wu, Haifeng Wang, Liang Huang, Qun Liu, Julia Ive, Wolfgang Macherey
{"title":"Findings of the Third Workshop on Automatic Simultaneous Translation","authors":"Ruiqing Zhang, Chuanqiang Zhang, Zhongjun He, Hua Wu, Haifeng Wang, Liang Huang, Qun Liu, Julia Ive, Wolfgang Macherey","doi":"10.18653/v1/2022.autosimtrans-1.1","DOIUrl":"https://doi.org/10.18653/v1/2022.autosimtrans-1.1","url":null,"abstract":"This paper reports the results of the shared task we hosted on the Third Workshop of Automatic Simultaneous Translation (AutoSimTrans). The shared task aims to promote the development of text-to-text and speech-to-text simultaneous translation, and includes Chinese-English and English-Spanish tracks. The number of systems submitted this year has increased fourfold compared with last year. Additionally, the top 1 ranked system in the speech-to-text track is the first end-to-end submission we have received in the past three years, which has shown great potential. This paper reports the results and descriptions of the 14 participating teams, compares different evaluation metrics, and revisits the ranking method.","PeriodicalId":444422,"journal":{"name":"Proceedings of the Third Workshop on Automatic Simultaneous Translation","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115154774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
End-to-End Simultaneous Speech Translation with Pretraining and Distillation: Huawei Noah’s System for AutoSimTranS 2022 基于预训练和蒸馏的端到端同步语音翻译:华为Noah的AutoSimTranS系统2022
Proceedings of the Third Workshop on Automatic Simultaneous Translation Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.autosimtrans-1.5
Xingshan Zeng, Pengfei Li, Liangyou Li, Qun Liu
{"title":"End-to-End Simultaneous Speech Translation with Pretraining and Distillation: Huawei Noah’s System for AutoSimTranS 2022","authors":"Xingshan Zeng, Pengfei Li, Liangyou Li, Qun Liu","doi":"10.18653/v1/2022.autosimtrans-1.5","DOIUrl":"https://doi.org/10.18653/v1/2022.autosimtrans-1.5","url":null,"abstract":"This paper describes the system submitted to AutoSimTrans 2022 from Huawei Noah’s Ark Lab, which won the first place in the audio input track of the Chinese-English translation task. Our system is based on RealTranS, an end-to-end simultaneous speech translation model. We enhance the model with pretraining, by initializing the acoustic encoder with ASR encoder, and the semantic encoder and decoder with NMT encoder and decoder, respectively. To relieve the data scarcity, we further construct pseudo training corpus as a kind of knowledge distillation with ASR data and the pretrained NMT model. Meanwhile, we also apply several techniques to improve the robustness and domain generalizability, including punctuation removal, token-level knowledge distillation and multi-domain finetuning. Experiments show that our system significantly outperforms the baselines at all latency and also verify the effectiveness of our proposed methods.","PeriodicalId":444422,"journal":{"name":"Proceedings of the Third Workshop on Automatic Simultaneous Translation","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114498198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
BIT-Xiaomi’s System for AutoSimTrans 2022 bit -小米的AutoSimTrans系统2022
Proceedings of the Third Workshop on Automatic Simultaneous Translation Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.autosimtrans-1.6
Mengge Liu, Xiang Li, Bao Chen, Yanzhi Tian, Tianwei Lan, Silin Li, Yuhang Guo, Jian Luan, Bin Wang
{"title":"BIT-Xiaomi’s System for AutoSimTrans 2022","authors":"Mengge Liu, Xiang Li, Bao Chen, Yanzhi Tian, Tianwei Lan, Silin Li, Yuhang Guo, Jian Luan, Bin Wang","doi":"10.18653/v1/2022.autosimtrans-1.6","DOIUrl":"https://doi.org/10.18653/v1/2022.autosimtrans-1.6","url":null,"abstract":"This system paper describes the BIT-Xiaomi simultaneous translation system for Autosimtrans 2022 simultaneous translation challenge. We participated in three tracks: the Zh-En text-to-text track, the Zh-En audio-to-text track and the En-Es test-to-text track. In our system, wait-k is employed to train prefix-to-prefix translation models. We integrate streaming chunking to detect boundaries as the source streaming read in. We further improve our system with data selection, data-augmentation and R-drop training methods. Results show that our wait-k implementation outperforms organizer’s baseline by 8 BLEU score at most, and our proposed streaming chunking method further improves about 2 BLEU in low latency regime.","PeriodicalId":444422,"journal":{"name":"Proceedings of the Third Workshop on Automatic Simultaneous Translation","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116564338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
System Description on Third Automatic Simultaneous Translation Workshop 第三届自动同声传译研讨会系统介绍
Proceedings of the Third Workshop on Automatic Simultaneous Translation Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.autosimtrans-1.4
Zhang Yiqiao
{"title":"System Description on Third Automatic Simultaneous Translation Workshop","authors":"Zhang Yiqiao","doi":"10.18653/v1/2022.autosimtrans-1.4","DOIUrl":"https://doi.org/10.18653/v1/2022.autosimtrans-1.4","url":null,"abstract":"This paper shows my submission to the Third Automatic Simultaneous Translation Workshop at NAACL2022.The submission includes Chinese audio to English text task, Chinese text to English text tast, and English text to Spanish text task.For the two text-to-text tasks, I use the STACL model of PaddleNLP. As for the audio-to-text task, I first use DeepSpeech2 to translate the audio into text, then apply the STACL model to handle the text-to-text task.The submission results show that the used method can get low delay with a few training samples.","PeriodicalId":444422,"journal":{"name":"Proceedings of the Third Workshop on Automatic Simultaneous Translation","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129163982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
System Description on Automatic Simultaneous Translation Workshop 自动同声传译车间系统描述
Proceedings of the Third Workshop on Automatic Simultaneous Translation Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.autosimtrans-1.3
Zecheng Li, Yue Sun, Haoze Li
{"title":"System Description on Automatic Simultaneous Translation Workshop","authors":"Zecheng Li, Yue Sun, Haoze Li","doi":"10.18653/v1/2022.autosimtrans-1.3","DOIUrl":"https://doi.org/10.18653/v1/2022.autosimtrans-1.3","url":null,"abstract":"This paper describes our system submitted on the third automatic simultaneous translation workshop at NAACL2022. We participate in the Chinese audio->English text direction of Chinese-to-English translation. Our speech-to-text system is a pipeline system, in which we resort to rhymological features for audio split, ASRT model for speech recoginition, STACL model for streaming text translation. To translate streaming text, we use wait-k policy trained to generate the target sentence concurrently with the source sentence, but always k words behind. We propose a competitive simultaneous translation system and rank 3rd in the audio input track. The code will release soon.","PeriodicalId":444422,"journal":{"name":"Proceedings of the Third Workshop on Automatic Simultaneous Translation","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122410147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信