Shuo Sun, H. Hou, Zongheng Yang, Yisong Wang, Nier Wu
{"title":"基于多奖励强化学习的低资源NMT对抗示例生成","authors":"Shuo Sun, H. Hou, Zongheng Yang, Yisong Wang, Nier Wu","doi":"10.1109/ICTAI56018.2022.00179","DOIUrl":null,"url":null,"abstract":"Weak robustness and noise adaptability are major issues for Low-Resource Neural Machine Translation (NMT) models. Adversarial example is currently a major tool to improve model robustness and how to generate an adversarial examples that can degrade the performance of the model and ensure semantic consistency is a challenging task. In this paper, we adopt multi-reward reinforcement learning to generate adversarial examples for low-resource NMT. Specifically, utilizing gradient ascent to modify the source sentence, the discriminator and changes estimate are used to determine whether the generated adversarial examples maintain semantic consistency and the overall modifications of adversarial examples. Furthermore, we also install a language model reward to measure the fluency of adversarial examples. Experimental results on low-resource translation tasks show that our method highly aggressive to the model while maintaining semantic constraints greatly. Moreover, the model performance is significantly improved after fine-tuning with adversarial examples.","PeriodicalId":354314,"journal":{"name":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Generating Adversarial Examples for Low-Resource NMT via Multi-Reward Reinforcement Learning\",\"authors\":\"Shuo Sun, H. Hou, Zongheng Yang, Yisong Wang, Nier Wu\",\"doi\":\"10.1109/ICTAI56018.2022.00179\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Weak robustness and noise adaptability are major issues for Low-Resource Neural Machine Translation (NMT) models. Adversarial example is currently a major tool to improve model robustness and how to generate an adversarial examples that can degrade the performance of the model and ensure semantic consistency is a challenging task. In this paper, we adopt multi-reward reinforcement learning to generate adversarial examples for low-resource NMT. Specifically, utilizing gradient ascent to modify the source sentence, the discriminator and changes estimate are used to determine whether the generated adversarial examples maintain semantic consistency and the overall modifications of adversarial examples. Furthermore, we also install a language model reward to measure the fluency of adversarial examples. Experimental results on low-resource translation tasks show that our method highly aggressive to the model while maintaining semantic constraints greatly. Moreover, the model performance is significantly improved after fine-tuning with adversarial examples.\",\"PeriodicalId\":354314,\"journal\":{\"name\":\"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICTAI56018.2022.00179\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI56018.2022.00179","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Generating Adversarial Examples for Low-Resource NMT via Multi-Reward Reinforcement Learning
Weak robustness and noise adaptability are major issues for Low-Resource Neural Machine Translation (NMT) models. Adversarial example is currently a major tool to improve model robustness and how to generate an adversarial examples that can degrade the performance of the model and ensure semantic consistency is a challenging task. In this paper, we adopt multi-reward reinforcement learning to generate adversarial examples for low-resource NMT. Specifically, utilizing gradient ascent to modify the source sentence, the discriminator and changes estimate are used to determine whether the generated adversarial examples maintain semantic consistency and the overall modifications of adversarial examples. Furthermore, we also install a language model reward to measure the fluency of adversarial examples. Experimental results on low-resource translation tasks show that our method highly aggressive to the model while maintaining semantic constraints greatly. Moreover, the model performance is significantly improved after fine-tuning with adversarial examples.