{"title":"时间插件:使用预训练图像去噪器进行无监督视频去噪","authors":"Zixuan Fu, Lanqing Guo, Chong Wang, Yufei Wang, Zhihao Li, Bihan Wen","doi":"arxiv-2409.11256","DOIUrl":null,"url":null,"abstract":"Recent advancements in deep learning have shown impressive results in image\nand video denoising, leveraging extensive pairs of noisy and noise-free data\nfor supervision. However, the challenge of acquiring paired videos for dynamic\nscenes hampers the practical deployment of deep video denoising techniques. In\ncontrast, this obstacle is less pronounced in image denoising, where paired\ndata is more readily available. Thus, a well-trained image denoiser could serve\nas a reliable spatial prior for video denoising. In this paper, we propose a\nnovel unsupervised video denoising framework, named ``Temporal As a Plugin''\n(TAP), which integrates tunable temporal modules into a pre-trained image\ndenoiser. By incorporating temporal modules, our method can harness temporal\ninformation across noisy frames, complementing its power of spatial denoising.\nFurthermore, we introduce a progressive fine-tuning strategy that refines each\ntemporal module using the generated pseudo clean video frames, progressively\nenhancing the network's denoising performance. Compared to other unsupervised\nvideo denoising methods, our framework demonstrates superior performance on\nboth sRGB and raw video denoising datasets.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Temporal As a Plugin: Unsupervised Video Denoising with Pre-Trained Image Denoisers\",\"authors\":\"Zixuan Fu, Lanqing Guo, Chong Wang, Yufei Wang, Zhihao Li, Bihan Wen\",\"doi\":\"arxiv-2409.11256\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent advancements in deep learning have shown impressive results in image\\nand video denoising, leveraging extensive pairs of noisy and noise-free data\\nfor supervision. However, the challenge of acquiring paired videos for dynamic\\nscenes hampers the practical deployment of deep video denoising techniques. In\\ncontrast, this obstacle is less pronounced in image denoising, where paired\\ndata is more readily available. Thus, a well-trained image denoiser could serve\\nas a reliable spatial prior for video denoising. In this paper, we propose a\\nnovel unsupervised video denoising framework, named ``Temporal As a Plugin''\\n(TAP), which integrates tunable temporal modules into a pre-trained image\\ndenoiser. By incorporating temporal modules, our method can harness temporal\\ninformation across noisy frames, complementing its power of spatial denoising.\\nFurthermore, we introduce a progressive fine-tuning strategy that refines each\\ntemporal module using the generated pseudo clean video frames, progressively\\nenhancing the network's denoising performance. Compared to other unsupervised\\nvideo denoising methods, our framework demonstrates superior performance on\\nboth sRGB and raw video denoising datasets.\",\"PeriodicalId\":501289,\"journal\":{\"name\":\"arXiv - EE - Image and Video Processing\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - EE - Image and Video Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11256\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Image and Video Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11256","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Temporal As a Plugin: Unsupervised Video Denoising with Pre-Trained Image Denoisers
Recent advancements in deep learning have shown impressive results in image
and video denoising, leveraging extensive pairs of noisy and noise-free data
for supervision. However, the challenge of acquiring paired videos for dynamic
scenes hampers the practical deployment of deep video denoising techniques. In
contrast, this obstacle is less pronounced in image denoising, where paired
data is more readily available. Thus, a well-trained image denoiser could serve
as a reliable spatial prior for video denoising. In this paper, we propose a
novel unsupervised video denoising framework, named ``Temporal As a Plugin''
(TAP), which integrates tunable temporal modules into a pre-trained image
denoiser. By incorporating temporal modules, our method can harness temporal
information across noisy frames, complementing its power of spatial denoising.
Furthermore, we introduce a progressive fine-tuning strategy that refines each
temporal module using the generated pseudo clean video frames, progressively
enhancing the network's denoising performance. Compared to other unsupervised
video denoising methods, our framework demonstrates superior performance on
both sRGB and raw video denoising datasets.