Songlin Lei , Qiuxia Yang , Ke Yang , Zhengpeng Zhao , Yuanyuan Pu
{"title":"Training-free style transfer via content-style image inversion","authors":"Songlin Lei , Qiuxia Yang , Ke Yang , Zhengpeng Zhao , Yuanyuan Pu","doi":"10.1016/j.cag.2025.104352","DOIUrl":null,"url":null,"abstract":"<div><div>Image style transfer aims to adapt a content image to a target style while preserving its structural information. Despite the strong generative capabilities of diffusion models, their application to style transfer faces two key challenges: (1) entangled content-style interplay during denoising, leading to suboptimal stylization, and (2) reliance on computationally expensive optimization (e.g., model fine-tuning or text inversion). To address these issues, we propose a training-free tri-path framework. The content and style paths separately leverage image inversion to extract content and style features, which are shared with the stylization path. Specifically, the content path preserves structure via residual connections and noised *h*-features, while the style path injects style through AdaIN-modulated self-attention features to avoid artifacts. Our method eliminates optimization overhead and ensures harmonious stylization by decoupling content-style control. Experiments demonstrate that our approach effectively retains content fidelity and style accuracy while minimizing artifacts.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"132 ","pages":"Article 104352"},"PeriodicalIF":2.8000,"publicationDate":"2025-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers & Graphics-Uk","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0097849325001931","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
Image style transfer aims to adapt a content image to a target style while preserving its structural information. Despite the strong generative capabilities of diffusion models, their application to style transfer faces two key challenges: (1) entangled content-style interplay during denoising, leading to suboptimal stylization, and (2) reliance on computationally expensive optimization (e.g., model fine-tuning or text inversion). To address these issues, we propose a training-free tri-path framework. The content and style paths separately leverage image inversion to extract content and style features, which are shared with the stylization path. Specifically, the content path preserves structure via residual connections and noised *h*-features, while the style path injects style through AdaIN-modulated self-attention features to avoid artifacts. Our method eliminates optimization overhead and ensures harmonious stylization by decoupling content-style control. Experiments demonstrate that our approach effectively retains content fidelity and style accuracy while minimizing artifacts.
期刊介绍:
Computers & Graphics is dedicated to disseminate information on research and applications of computer graphics (CG) techniques. The journal encourages articles on:
1. Research and applications of interactive computer graphics. We are particularly interested in novel interaction techniques and applications of CG to problem domains.
2. State-of-the-art papers on late-breaking, cutting-edge research on CG.
3. Information on innovative uses of graphics principles and technologies.
4. Tutorial papers on both teaching CG principles and innovative uses of CG in education.