Li Li , Xinpeng Zhang , Guorui Feng , Zichi Wang , Deyang Wu , Hanzhou Wu
{"title":"Robust watermarking for diffusion models based on STDM and latent space fine-tuning","authors":"Li Li , Xinpeng Zhang , Guorui Feng , Zichi Wang , Deyang Wu , Hanzhou Wu","doi":"10.1016/j.jisa.2025.104167","DOIUrl":null,"url":null,"abstract":"<div><div>Diffusion models (DMs) have demonstrated remarkable capabilities in generating high-quality images, but their potential for disseminating harmful misinformation raises significant concerns. Although reversible watermarking techniques can trace AI-generated images to their source models by embedding watermarks in the latent space, existing methods suffer from two critical drawbacks: (i) limited embedding capacity hinders unique model identification, and (ii) information loss during latent-space re-encoding compromises robustness, exacerbating the inherent trade-off between capacity and robustness. To address these limitations, we propose a novel watermarking framework based on Spread Transform Dither Modulation (STDM) that embeds watermarks into intermediate latent vectors during the diffusion process. Our approach operates in three key steps: (i) executing the standard diffusion process to obtain an intermediate latent vector, (ii) embedding watermarks into the mid-frequency DCT coefficients of this vector using ring-shaped STDM modulation, and (iii) completing the diffusion process to generate the final watermarked image. For watermark extraction, we employ a finely tuned VAE encoder to map the image back to latent space, followed by DDIM inversion and STDM-based extraction. Furthermore, we introduce a joint fine-tuning strategy that optimizes both the encoder and decoder of the diffusion model using watermarked latent vectors, significantly enhancing robustness. Experimental results demonstrate that our method achieves a maximum watermark embedding capacity of 256 bits while maintaining a high extraction accuracy of 98%. The proposed approach exhibits remarkable robustness against various attacks, with significant improvements over baseline methods.</div></div>","PeriodicalId":48638,"journal":{"name":"Journal of Information Security and Applications","volume":"93 ","pages":"Article 104167"},"PeriodicalIF":3.8000,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Information Security and Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2214212625002042","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Diffusion models (DMs) have demonstrated remarkable capabilities in generating high-quality images, but their potential for disseminating harmful misinformation raises significant concerns. Although reversible watermarking techniques can trace AI-generated images to their source models by embedding watermarks in the latent space, existing methods suffer from two critical drawbacks: (i) limited embedding capacity hinders unique model identification, and (ii) information loss during latent-space re-encoding compromises robustness, exacerbating the inherent trade-off between capacity and robustness. To address these limitations, we propose a novel watermarking framework based on Spread Transform Dither Modulation (STDM) that embeds watermarks into intermediate latent vectors during the diffusion process. Our approach operates in three key steps: (i) executing the standard diffusion process to obtain an intermediate latent vector, (ii) embedding watermarks into the mid-frequency DCT coefficients of this vector using ring-shaped STDM modulation, and (iii) completing the diffusion process to generate the final watermarked image. For watermark extraction, we employ a finely tuned VAE encoder to map the image back to latent space, followed by DDIM inversion and STDM-based extraction. Furthermore, we introduce a joint fine-tuning strategy that optimizes both the encoder and decoder of the diffusion model using watermarked latent vectors, significantly enhancing robustness. Experimental results demonstrate that our method achieves a maximum watermark embedding capacity of 256 bits while maintaining a high extraction accuracy of 98%. The proposed approach exhibits remarkable robustness against various attacks, with significant improvements over baseline methods.
期刊介绍:
Journal of Information Security and Applications (JISA) focuses on the original research and practice-driven applications with relevance to information security and applications. JISA provides a common linkage between a vibrant scientific and research community and industry professionals by offering a clear view on modern problems and challenges in information security, as well as identifying promising scientific and "best-practice" solutions. JISA issues offer a balance between original research work and innovative industrial approaches by internationally renowned information security experts and researchers.