Yuxin Gong, Shen Wang, Xunzhi Jiang, Tingyue Yu, Fanghui Sun
{"title":"使用随机编码网络的对抗性净化","authors":"Yuxin Gong, Shen Wang, Xunzhi Jiang, Tingyue Yu, Fanghui Sun","doi":"10.1016/j.asoc.2025.113604","DOIUrl":null,"url":null,"abstract":"<div><div>Deep neural networks (DNNs) have revealed vulnerabilities to adversarial examples, which can deceive models with high confidence. This has given rise to serious threats in security-critical domains. Adversarial defense methods have been extensively studied to counter adversarial attacks. Adversarial purification, as a major defense strategy, attempts to recover adversarial examples to clean counterparts by filtering out perturbations. However, many purification defenses struggle against white-box attacks where the target and defense models are known. Additionally, the training processes against specific attacks can compromise models’ adaptability to unknown attacks, and purification operations may destroy key features of inputs. In this paper, we propose the random encoding network (REN), which consists of a random encoding denoiser and a diverse classifier to enhance the robustness of adversarial purification defense models. The internal part of the denoiser leverages adversarial sparse coding to purify examples by filtering out perturbations and noise as much as possible while preserving critical features of inputs. The external part of the denoiser employs a dynamic random mechanism to implement random encoding, thereby enhancing the models’ uncertainty. Moreover, the classifier is subjected to a diversity constraint to promote variation among random sub-models. Experimental results demonstrate that REN exhibits strong defensive generalization capabilities, effectively countering adversarial examples across diverse attack types and settings. For the CIFAR-10 and SVHN datasets, the clean-trained REN achieves average adversarial accuracies of 63.26% and 59.78% against white-box attacks, while the adversarial-trained REN achieves 68.27% and 72.39%, respectively. When faced with unknown attack scenarios, REN is more effective than state-of-the-art defense methods.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"183 ","pages":"Article 113604"},"PeriodicalIF":7.2000,"publicationDate":"2025-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Adversarial purification using random encoding networks\",\"authors\":\"Yuxin Gong, Shen Wang, Xunzhi Jiang, Tingyue Yu, Fanghui Sun\",\"doi\":\"10.1016/j.asoc.2025.113604\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Deep neural networks (DNNs) have revealed vulnerabilities to adversarial examples, which can deceive models with high confidence. This has given rise to serious threats in security-critical domains. Adversarial defense methods have been extensively studied to counter adversarial attacks. Adversarial purification, as a major defense strategy, attempts to recover adversarial examples to clean counterparts by filtering out perturbations. However, many purification defenses struggle against white-box attacks where the target and defense models are known. Additionally, the training processes against specific attacks can compromise models’ adaptability to unknown attacks, and purification operations may destroy key features of inputs. In this paper, we propose the random encoding network (REN), which consists of a random encoding denoiser and a diverse classifier to enhance the robustness of adversarial purification defense models. The internal part of the denoiser leverages adversarial sparse coding to purify examples by filtering out perturbations and noise as much as possible while preserving critical features of inputs. The external part of the denoiser employs a dynamic random mechanism to implement random encoding, thereby enhancing the models’ uncertainty. Moreover, the classifier is subjected to a diversity constraint to promote variation among random sub-models. Experimental results demonstrate that REN exhibits strong defensive generalization capabilities, effectively countering adversarial examples across diverse attack types and settings. For the CIFAR-10 and SVHN datasets, the clean-trained REN achieves average adversarial accuracies of 63.26% and 59.78% against white-box attacks, while the adversarial-trained REN achieves 68.27% and 72.39%, respectively. When faced with unknown attack scenarios, REN is more effective than state-of-the-art defense methods.</div></div>\",\"PeriodicalId\":50737,\"journal\":{\"name\":\"Applied Soft Computing\",\"volume\":\"183 \",\"pages\":\"Article 113604\"},\"PeriodicalIF\":7.2000,\"publicationDate\":\"2025-07-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Soft Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1568494625009159\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494625009159","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Adversarial purification using random encoding networks
Deep neural networks (DNNs) have revealed vulnerabilities to adversarial examples, which can deceive models with high confidence. This has given rise to serious threats in security-critical domains. Adversarial defense methods have been extensively studied to counter adversarial attacks. Adversarial purification, as a major defense strategy, attempts to recover adversarial examples to clean counterparts by filtering out perturbations. However, many purification defenses struggle against white-box attacks where the target and defense models are known. Additionally, the training processes against specific attacks can compromise models’ adaptability to unknown attacks, and purification operations may destroy key features of inputs. In this paper, we propose the random encoding network (REN), which consists of a random encoding denoiser and a diverse classifier to enhance the robustness of adversarial purification defense models. The internal part of the denoiser leverages adversarial sparse coding to purify examples by filtering out perturbations and noise as much as possible while preserving critical features of inputs. The external part of the denoiser employs a dynamic random mechanism to implement random encoding, thereby enhancing the models’ uncertainty. Moreover, the classifier is subjected to a diversity constraint to promote variation among random sub-models. Experimental results demonstrate that REN exhibits strong defensive generalization capabilities, effectively countering adversarial examples across diverse attack types and settings. For the CIFAR-10 and SVHN datasets, the clean-trained REN achieves average adversarial accuracies of 63.26% and 59.78% against white-box attacks, while the adversarial-trained REN achieves 68.27% and 72.39%, respectively. When faced with unknown attack scenarios, REN is more effective than state-of-the-art defense methods.
期刊介绍:
Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities.
Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.