基于解纠缠表示的无监督对抗性扰动消除

Lingyun Jiang, Kai Qiao, Ruoxi Qin, Jian Chen, Haibing Bu, Bin Yan
{"title":"基于解纠缠表示的无监督对抗性扰动消除","authors":"Lingyun Jiang, Kai Qiao, Ruoxi Qin, Jian Chen, Haibing Bu, Bin Yan","doi":"10.1145/3351917.3351987","DOIUrl":null,"url":null,"abstract":"Although deep neural networks (DNNs) could achieve state-of-the-art performance while recognizing images, they often vulnerable to adversarial examples where input intended to be added the small magnitude perturbations may mislead them to incorrect results. It is worth researching on defending against adversarial examples due to the potential security threats. In this paper, we propose an unsupervised method for eliminating adversarial perturbation based on disentangled representations. To achieve adversarial defense, we propose extracting the content and perturbation features of adversarial examples by content encoders and perturbation encoders. Meanwhile, to handle the unpaired training data, we introduce a cross-cycle consistency loss based on disentangled representations and a perturbation branch. We also add an adversarial loss on recovered images to make DNNs predict right. Qualitative results show that our method can eliminate adversarial perturbation without paired training data. We perform extensive experiments on two public datasets MNIST and CIFAR10, which is shown the efficiency of resisting adversarial examples.","PeriodicalId":367885,"journal":{"name":"Proceedings of the 2019 4th International Conference on Automation, Control and Robotics Engineering","volume":"58 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Unsupervised Adversarial Perturbation Eliminating via Disentangled Representations\",\"authors\":\"Lingyun Jiang, Kai Qiao, Ruoxi Qin, Jian Chen, Haibing Bu, Bin Yan\",\"doi\":\"10.1145/3351917.3351987\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Although deep neural networks (DNNs) could achieve state-of-the-art performance while recognizing images, they often vulnerable to adversarial examples where input intended to be added the small magnitude perturbations may mislead them to incorrect results. It is worth researching on defending against adversarial examples due to the potential security threats. In this paper, we propose an unsupervised method for eliminating adversarial perturbation based on disentangled representations. To achieve adversarial defense, we propose extracting the content and perturbation features of adversarial examples by content encoders and perturbation encoders. Meanwhile, to handle the unpaired training data, we introduce a cross-cycle consistency loss based on disentangled representations and a perturbation branch. We also add an adversarial loss on recovered images to make DNNs predict right. Qualitative results show that our method can eliminate adversarial perturbation without paired training data. We perform extensive experiments on two public datasets MNIST and CIFAR10, which is shown the efficiency of resisting adversarial examples.\",\"PeriodicalId\":367885,\"journal\":{\"name\":\"Proceedings of the 2019 4th International Conference on Automation, Control and Robotics Engineering\",\"volume\":\"58 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-07-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2019 4th International Conference on Automation, Control and Robotics Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3351917.3351987\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2019 4th International Conference on Automation, Control and Robotics Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3351917.3351987","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

虽然深度神经网络(dnn)在识别图像时可以达到最先进的性能,但它们往往容易受到对抗性示例的影响,其中意图添加小幅度扰动的输入可能会误导它们获得不正确的结果。由于潜在的安全威胁,对抗性示例的防御是值得研究的。在本文中,我们提出了一种基于解纠缠表示的消除对抗性扰动的无监督方法。为了实现对抗性防御,我们提出了通过内容编码器和摄动编码器提取对抗性样本的内容和摄动特征。同时,为了处理未配对的训练数据,我们引入了基于解纠缠表示和扰动分支的交叉周期一致性损失。我们还在恢复的图像上添加了对抗损失,以使dnn预测正确。定性结果表明,该方法可以在不需要配对训练数据的情况下消除对抗性扰动。我们在两个公共数据集MNIST和CIFAR10上进行了大量的实验,证明了抵抗对抗样本的效率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Unsupervised Adversarial Perturbation Eliminating via Disentangled Representations
Although deep neural networks (DNNs) could achieve state-of-the-art performance while recognizing images, they often vulnerable to adversarial examples where input intended to be added the small magnitude perturbations may mislead them to incorrect results. It is worth researching on defending against adversarial examples due to the potential security threats. In this paper, we propose an unsupervised method for eliminating adversarial perturbation based on disentangled representations. To achieve adversarial defense, we propose extracting the content and perturbation features of adversarial examples by content encoders and perturbation encoders. Meanwhile, to handle the unpaired training data, we introduce a cross-cycle consistency loss based on disentangled representations and a perturbation branch. We also add an adversarial loss on recovered images to make DNNs predict right. Qualitative results show that our method can eliminate adversarial perturbation without paired training data. We perform extensive experiments on two public datasets MNIST and CIFAR10, which is shown the efficiency of resisting adversarial examples.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信