{"title":"Towards Effective User Attribution for Latent Diffusion Models via Watermark-Informed Blending","authors":"Yongyang Pan, Xiaohong Liu, Siqi Luo, Yi Xin, Xiao Guo, Xiaoming Liu, Xiongkuo Min, Guangtao Zhai","doi":"arxiv-2409.10958","DOIUrl":null,"url":null,"abstract":"Rapid advancements in multimodal large language models have enabled the\ncreation of hyper-realistic images from textual descriptions. However, these\nadvancements also raise significant concerns about unauthorized use, which\nhinders their broader distribution. Traditional watermarking methods often\nrequire complex integration or degrade image quality. To address these\nchallenges, we introduce a novel framework Towards Effective user Attribution\nfor latent diffusion models via Watermark-Informed Blending (TEAWIB). TEAWIB\nincorporates a unique ready-to-use configuration approach that allows seamless\nintegration of user-specific watermarks into generative models. This approach\nensures that each user can directly apply a pre-configured set of parameters to\nthe model without altering the original model parameters or compromising image\nquality. Additionally, noise and augmentation operations are embedded at the\npixel level to further secure and stabilize watermarked images. Extensive\nexperiments validate the effectiveness of TEAWIB, showcasing the\nstate-of-the-art performance in perceptual quality and attribution accuracy.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Image and Video Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10958","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Rapid advancements in multimodal large language models have enabled the
creation of hyper-realistic images from textual descriptions. However, these
advancements also raise significant concerns about unauthorized use, which
hinders their broader distribution. Traditional watermarking methods often
require complex integration or degrade image quality. To address these
challenges, we introduce a novel framework Towards Effective user Attribution
for latent diffusion models via Watermark-Informed Blending (TEAWIB). TEAWIB
incorporates a unique ready-to-use configuration approach that allows seamless
integration of user-specific watermarks into generative models. This approach
ensures that each user can directly apply a pre-configured set of parameters to
the model without altering the original model parameters or compromising image
quality. Additionally, noise and augmentation operations are embedded at the
pixel level to further secure and stabilize watermarked images. Extensive
experiments validate the effectiveness of TEAWIB, showcasing the
state-of-the-art performance in perceptual quality and attribution accuracy.