Junjian Li , Honglong Chen , Yudong Gao , Shaozhong Guo , Kai Lin , Yuping Liu , Peng Sun
{"title":"BABE: Backdoor attack with bokeh effects via latent separation suppression","authors":"Junjian Li , Honglong Chen , Yudong Gao , Shaozhong Guo , Kai Lin , Yuping Liu , Peng Sun","doi":"10.1016/j.engappai.2024.109462","DOIUrl":null,"url":null,"abstract":"<div><div>The escalating menace of backdoor attacks constitutes a formidable obstacle to the ongoing advancement of deep neural networks (DNNs), particularly in the security-sensitive applications such as face recognition and self-driving. Backdoored models render deliberately incorrect predictions on the inputs with the crafted triggers while behaving normally with the benign ones. Despite demonstrating the varying degrees of threat, existing backdoor attack strategies often prioritize stealthiness and defense evasions but neglect the practical feasibility in the real-world deployment scenarios. In this paper, we develop a backdoor attack leveraging bokeh effects (<span><math><mrow><mi>B</mi><mi>A</mi><mi>B</mi><mi>E</mi></mrow></math></span>), which introduces the bokeh effects as the trigger. Once the backdoored model is deployed in the vision application, the model’s malicious behaviors can be activated only by utilizing the captured bokeh images without any other modifications. Specially, we employ the saliency and depth estimation maps to derive the bokeh images, thereby serving as the poisoned samples. Furthermore, to avoid the latent separation of the generated poisoned images, we propose distinct attack strategies on the basis of the adversary’s prior abilities. For the adversary only with the data manipulation, we retain the original semantic labels for a subset of poisoned data during the training process. For the adversary with the manipulation of both the data and models, we construct a reference model trained on the clean samples to impose constraints on the latent representations of the poisoned images. Extensive experiments demonstrate the attack effects of the proposed <span><math><mrow><mi>B</mi><mi>A</mi><mi>B</mi><mi>E</mi></mrow></math></span>, even on the bokeh photos captured from Digital Still Cameras (DSC) and smartphones.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":7.5000,"publicationDate":"2024-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197624016208","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The escalating menace of backdoor attacks constitutes a formidable obstacle to the ongoing advancement of deep neural networks (DNNs), particularly in the security-sensitive applications such as face recognition and self-driving. Backdoored models render deliberately incorrect predictions on the inputs with the crafted triggers while behaving normally with the benign ones. Despite demonstrating the varying degrees of threat, existing backdoor attack strategies often prioritize stealthiness and defense evasions but neglect the practical feasibility in the real-world deployment scenarios. In this paper, we develop a backdoor attack leveraging bokeh effects (), which introduces the bokeh effects as the trigger. Once the backdoored model is deployed in the vision application, the model’s malicious behaviors can be activated only by utilizing the captured bokeh images without any other modifications. Specially, we employ the saliency and depth estimation maps to derive the bokeh images, thereby serving as the poisoned samples. Furthermore, to avoid the latent separation of the generated poisoned images, we propose distinct attack strategies on the basis of the adversary’s prior abilities. For the adversary only with the data manipulation, we retain the original semantic labels for a subset of poisoned data during the training process. For the adversary with the manipulation of both the data and models, we construct a reference model trained on the clean samples to impose constraints on the latent representations of the poisoned images. Extensive experiments demonstrate the attack effects of the proposed , even on the bokeh photos captured from Digital Still Cameras (DSC) and smartphones.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.