Wenhui Li , Bo Li , Weizhi Nie , Lanjun Wang , An-An Liu
{"title":"跨模态对抗性攻击中最优目标代码引导的多元摄动","authors":"Wenhui Li , Bo Li , Weizhi Nie , Lanjun Wang , An-An Liu","doi":"10.1016/j.ipm.2025.104214","DOIUrl":null,"url":null,"abstract":"<div><div>Cross-modal retrieval models are vulnerable to adversarial samples, thus exploring efficient attack methods can help researchers understand the essence of adversarial attack, evaluate the robustness of models, and promote the development of more reliable models. Although existing adversarial attack methods have achieved promising results, how to further improve the transferability of adversarial examples remains an open question. In this paper, we propose a novel transferable targeted attack method. First, we introduce an optimal target code optimization strategy to obtain representative target codes. Subsequently, when generating adversarial examples, we propose a random perturbation strategy to diversify the potential input patterns of perturbations by introducing randomness, thus automatically enhancing the generalization of samples. overcoming the limitations of existing methods that depend on specific image augmentation techniques. Experiments show that this framework can generate highly transferable adversarial samples, for example, when transferring attacks from VGG-F to ResNet50, the proposed method outperforms the SOTA by 10.41% in the I2T task on the MS-COCO dataset; in addition, samples generated by small models can also successfully attack large models, which will help researchers study the existence of adversarial attacks from a new perspective.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"62 5","pages":"Article 104214"},"PeriodicalIF":7.4000,"publicationDate":"2025-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Diversified perturbation guided by optimal target code for cross-modal adversarial attack\",\"authors\":\"Wenhui Li , Bo Li , Weizhi Nie , Lanjun Wang , An-An Liu\",\"doi\":\"10.1016/j.ipm.2025.104214\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Cross-modal retrieval models are vulnerable to adversarial samples, thus exploring efficient attack methods can help researchers understand the essence of adversarial attack, evaluate the robustness of models, and promote the development of more reliable models. Although existing adversarial attack methods have achieved promising results, how to further improve the transferability of adversarial examples remains an open question. In this paper, we propose a novel transferable targeted attack method. First, we introduce an optimal target code optimization strategy to obtain representative target codes. Subsequently, when generating adversarial examples, we propose a random perturbation strategy to diversify the potential input patterns of perturbations by introducing randomness, thus automatically enhancing the generalization of samples. overcoming the limitations of existing methods that depend on specific image augmentation techniques. Experiments show that this framework can generate highly transferable adversarial samples, for example, when transferring attacks from VGG-F to ResNet50, the proposed method outperforms the SOTA by 10.41% in the I2T task on the MS-COCO dataset; in addition, samples generated by small models can also successfully attack large models, which will help researchers study the existence of adversarial attacks from a new perspective.</div></div>\",\"PeriodicalId\":50365,\"journal\":{\"name\":\"Information Processing & Management\",\"volume\":\"62 5\",\"pages\":\"Article 104214\"},\"PeriodicalIF\":7.4000,\"publicationDate\":\"2025-05-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Processing & Management\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306457325001554\",\"RegionNum\":1,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325001554","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Diversified perturbation guided by optimal target code for cross-modal adversarial attack
Cross-modal retrieval models are vulnerable to adversarial samples, thus exploring efficient attack methods can help researchers understand the essence of adversarial attack, evaluate the robustness of models, and promote the development of more reliable models. Although existing adversarial attack methods have achieved promising results, how to further improve the transferability of adversarial examples remains an open question. In this paper, we propose a novel transferable targeted attack method. First, we introduce an optimal target code optimization strategy to obtain representative target codes. Subsequently, when generating adversarial examples, we propose a random perturbation strategy to diversify the potential input patterns of perturbations by introducing randomness, thus automatically enhancing the generalization of samples. overcoming the limitations of existing methods that depend on specific image augmentation techniques. Experiments show that this framework can generate highly transferable adversarial samples, for example, when transferring attacks from VGG-F to ResNet50, the proposed method outperforms the SOTA by 10.41% in the I2T task on the MS-COCO dataset; in addition, samples generated by small models can also successfully attack large models, which will help researchers study the existence of adversarial attacks from a new perspective.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.