Yang Luo, Yiheng Zhang, Zhaofan Qiu, Ting Yao, Zhineng Chen, Yu-Gang Jiang, Tao Mei
{"title":"FreeEnhance:通过内容一致的噪声和去噪过程实现无调谐图像增强","authors":"Yang Luo, Yiheng Zhang, Zhaofan Qiu, Ting Yao, Zhineng Chen, Yu-Gang Jiang, Tao Mei","doi":"arxiv-2409.07451","DOIUrl":null,"url":null,"abstract":"The emergence of text-to-image generation models has led to the recognition\nthat image enhancement, performed as post-processing, would significantly\nimprove the visual quality of the generated images. Exploring diffusion models\nto enhance the generated images nevertheless is not trivial and necessitates to\ndelicately enrich plentiful details while preserving the visual appearance of\nkey content in the original image. In this paper, we propose a novel framework,\nnamely FreeEnhance, for content-consistent image enhancement using the\noff-the-shelf image diffusion models. Technically, FreeEnhance is a two-stage\nprocess that firstly adds random noise to the input image and then capitalizes\non a pre-trained image diffusion model (i.e., Latent Diffusion Models) to\ndenoise and enhance the image details. In the noising stage, FreeEnhance is\ndevised to add lighter noise to the region with higher frequency to preserve\nthe high-frequent patterns (e.g., edge, corner) in the original image. In the\ndenoising stage, we present three target properties as constraints to\nregularize the predicted noise, enhancing images with high acutance and high\nvisual quality. Extensive experiments conducted on the HPDv2 dataset\ndemonstrate that our FreeEnhance outperforms the state-of-the-art image\nenhancement models in terms of quantitative metrics and human preference. More\nremarkably, FreeEnhance also shows higher human preference compared to the\ncommercial image enhancement solution of Magnific AI.","PeriodicalId":501480,"journal":{"name":"arXiv - CS - Multimedia","volume":"13 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FreeEnhance: Tuning-Free Image Enhancement via Content-Consistent Noising-and-Denoising Process\",\"authors\":\"Yang Luo, Yiheng Zhang, Zhaofan Qiu, Ting Yao, Zhineng Chen, Yu-Gang Jiang, Tao Mei\",\"doi\":\"arxiv-2409.07451\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The emergence of text-to-image generation models has led to the recognition\\nthat image enhancement, performed as post-processing, would significantly\\nimprove the visual quality of the generated images. Exploring diffusion models\\nto enhance the generated images nevertheless is not trivial and necessitates to\\ndelicately enrich plentiful details while preserving the visual appearance of\\nkey content in the original image. In this paper, we propose a novel framework,\\nnamely FreeEnhance, for content-consistent image enhancement using the\\noff-the-shelf image diffusion models. Technically, FreeEnhance is a two-stage\\nprocess that firstly adds random noise to the input image and then capitalizes\\non a pre-trained image diffusion model (i.e., Latent Diffusion Models) to\\ndenoise and enhance the image details. In the noising stage, FreeEnhance is\\ndevised to add lighter noise to the region with higher frequency to preserve\\nthe high-frequent patterns (e.g., edge, corner) in the original image. In the\\ndenoising stage, we present three target properties as constraints to\\nregularize the predicted noise, enhancing images with high acutance and high\\nvisual quality. Extensive experiments conducted on the HPDv2 dataset\\ndemonstrate that our FreeEnhance outperforms the state-of-the-art image\\nenhancement models in terms of quantitative metrics and human preference. More\\nremarkably, FreeEnhance also shows higher human preference compared to the\\ncommercial image enhancement solution of Magnific AI.\",\"PeriodicalId\":501480,\"journal\":{\"name\":\"arXiv - CS - Multimedia\",\"volume\":\"13 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Multimedia\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.07451\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multimedia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07451","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
FreeEnhance: Tuning-Free Image Enhancement via Content-Consistent Noising-and-Denoising Process
The emergence of text-to-image generation models has led to the recognition
that image enhancement, performed as post-processing, would significantly
improve the visual quality of the generated images. Exploring diffusion models
to enhance the generated images nevertheless is not trivial and necessitates to
delicately enrich plentiful details while preserving the visual appearance of
key content in the original image. In this paper, we propose a novel framework,
namely FreeEnhance, for content-consistent image enhancement using the
off-the-shelf image diffusion models. Technically, FreeEnhance is a two-stage
process that firstly adds random noise to the input image and then capitalizes
on a pre-trained image diffusion model (i.e., Latent Diffusion Models) to
denoise and enhance the image details. In the noising stage, FreeEnhance is
devised to add lighter noise to the region with higher frequency to preserve
the high-frequent patterns (e.g., edge, corner) in the original image. In the
denoising stage, we present three target properties as constraints to
regularize the predicted noise, enhancing images with high acutance and high
visual quality. Extensive experiments conducted on the HPDv2 dataset
demonstrate that our FreeEnhance outperforms the state-of-the-art image
enhancement models in terms of quantitative metrics and human preference. More
remarkably, FreeEnhance also shows higher human preference compared to the
commercial image enhancement solution of Magnific AI.