Zhenghan Gao , Chengming Liu , Yucheng Shi , Xin Guo , Jing Xu , Hong Zhang , Lei Shi
{"title":"FTA2C: Achieving superior trade-off between accuracy and robustness in adversarial training","authors":"Zhenghan Gao , Chengming Liu , Yucheng Shi , Xin Guo , Jing Xu , Hong Zhang , Lei Shi","doi":"10.1016/j.neunet.2025.108176","DOIUrl":null,"url":null,"abstract":"<div><div>Deep neural networks are notoriously vulnerable to adversarial perturbations, largely due to the presence of non-robust features that destabilize model performance. Traditional Adversarial Training (AT) methods on feature space typically operate on one part of features individually, resulting in the loss of useful information in them, and improve robustness at the expense of accuracy, making it difficult to optimize the inherent trade-off between the two. To address this challenge, we propose a novel plug-in method termed Feature Transformation Alignment and Compression (FTA2C). FTA2C comprises three key components. First, a compression network constrains the perturbation space to reduce the vulnerability of non-robust features. Second, a feature transformation network enhances the expressiveness of robust features. Third, an alignment mechanism enforces consistency between adversarial and natural samples in the robust feature space. The above mechanism achieves co-processing of the two parts of the feature. Additionally, we propose the Defense Efficiency Metric (DEM) to evaluate defense methods. DEM quantifies the trade-off between maintaining natural accuracy and enhancing adversarial robustness, offering a unified and interpretable standard for comparing defense strategies. Extensive experiments conducted on four benchmark datasets demonstrate that FTA2C significantly improvements robustness under the high-level accuracy, resulting in superior trade-off performance. Our code is available at <span><span>https://github.com/HymanGao31/FTA2C</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"194 ","pages":"Article 108176"},"PeriodicalIF":6.3000,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025010561","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Deep neural networks are notoriously vulnerable to adversarial perturbations, largely due to the presence of non-robust features that destabilize model performance. Traditional Adversarial Training (AT) methods on feature space typically operate on one part of features individually, resulting in the loss of useful information in them, and improve robustness at the expense of accuracy, making it difficult to optimize the inherent trade-off between the two. To address this challenge, we propose a novel plug-in method termed Feature Transformation Alignment and Compression (FTA2C). FTA2C comprises three key components. First, a compression network constrains the perturbation space to reduce the vulnerability of non-robust features. Second, a feature transformation network enhances the expressiveness of robust features. Third, an alignment mechanism enforces consistency between adversarial and natural samples in the robust feature space. The above mechanism achieves co-processing of the two parts of the feature. Additionally, we propose the Defense Efficiency Metric (DEM) to evaluate defense methods. DEM quantifies the trade-off between maintaining natural accuracy and enhancing adversarial robustness, offering a unified and interpretable standard for comparing defense strategies. Extensive experiments conducted on four benchmark datasets demonstrate that FTA2C significantly improvements robustness under the high-level accuracy, resulting in superior trade-off performance. Our code is available at https://github.com/HymanGao31/FTA2C.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.