Shihui Zhang , Shaojie Han , Sheng Yang , Xueqiang Han , Junbin Su , Gangzheng Zhai , Houlin Wang
{"title":"Transferable targeted adversarial attack via multi-source perturbation generation and integration","authors":"Shihui Zhang , Shaojie Han , Sheng Yang , Xueqiang Han , Junbin Su , Gangzheng Zhai , Houlin Wang","doi":"10.1016/j.jvcir.2025.104493","DOIUrl":null,"url":null,"abstract":"<div><div>With the rapid development of artificial intelligence, deep learning models have been applied in the field of society (e.g., video or image representation). However, due to the presence of adversarial examples, these models exhibit obvious fragility, which has become a major challenge restricting society development. Therefore, studying the generation process and achieving high transferability of adversarial examples are of utmost importance. In this paper, we propose a transferable targeted adversarial attack method called Multi-source Perturbation Generation and Integration (MPGI) to address the vulnerability and uncertainty of deep learning models. Specifically, MPGI consists of three critical designs to achieve targeted transferability of adversarial examples. Firstly, we propose a Collaborative Feature Fusion (CFF) component, which reduces the impact of original example feature on model classification by considering collaboration in feature fusion. Subsequently, we propose a Multi-scale Perturbation Dynamic Fusion (MPDF) module to fuse perturbations from different scales for enriching perturbation diversity. Finally, we innovatively investigate a novel Logit Margin with Penalty (LMP) loss to further enhance the misleading ability of the examples. The LMP, as a pluggable part, offers the potential to be leveraged by different approaches for boosting performance. In summary, MPGI can effectively achieve targeted attacks, expose the shortcomings of existing models, and promote the development of artificial intelligence in terms of security. Extensive experiments on ImageNet-Compatible and CIFAR-10 datasets demonstrate the superiority of the proposed method. For instance, the attack success rate increases by 17.6% and 17.0% compared to state-of-the-art method when transferred from DN-121 to Inc-v3 and MB-v2 models.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"111 ","pages":"Article 104493"},"PeriodicalIF":3.1000,"publicationDate":"2025-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Visual Communication and Image Representation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1047320325001075","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
With the rapid development of artificial intelligence, deep learning models have been applied in the field of society (e.g., video or image representation). However, due to the presence of adversarial examples, these models exhibit obvious fragility, which has become a major challenge restricting society development. Therefore, studying the generation process and achieving high transferability of adversarial examples are of utmost importance. In this paper, we propose a transferable targeted adversarial attack method called Multi-source Perturbation Generation and Integration (MPGI) to address the vulnerability and uncertainty of deep learning models. Specifically, MPGI consists of three critical designs to achieve targeted transferability of adversarial examples. Firstly, we propose a Collaborative Feature Fusion (CFF) component, which reduces the impact of original example feature on model classification by considering collaboration in feature fusion. Subsequently, we propose a Multi-scale Perturbation Dynamic Fusion (MPDF) module to fuse perturbations from different scales for enriching perturbation diversity. Finally, we innovatively investigate a novel Logit Margin with Penalty (LMP) loss to further enhance the misleading ability of the examples. The LMP, as a pluggable part, offers the potential to be leveraged by different approaches for boosting performance. In summary, MPGI can effectively achieve targeted attacks, expose the shortcomings of existing models, and promote the development of artificial intelligence in terms of security. Extensive experiments on ImageNet-Compatible and CIFAR-10 datasets demonstrate the superiority of the proposed method. For instance, the attack success rate increases by 17.6% and 17.0% compared to state-of-the-art method when transferred from DN-121 to Inc-v3 and MB-v2 models.
期刊介绍:
The Journal of Visual Communication and Image Representation publishes papers on state-of-the-art visual communication and image representation, with emphasis on novel technologies and theoretical work in this multidisciplinary area of pure and applied research. The field of visual communication and image representation is considered in its broadest sense and covers both digital and analog aspects as well as processing and communication in biological visual systems.