Xianfeng Tang, Peining Zhen, M. Kang, Hang Yi, Wei Wang, Hai-Bao Chen
{"title":"学习卷积神经网络视频去噪的丰富特征","authors":"Xianfeng Tang, Peining Zhen, M. Kang, Hang Yi, Wei Wang, Hai-Bao Chen","doi":"10.1109/APCCAS50809.2020.9301660","DOIUrl":null,"url":null,"abstract":"Video denoising is of great significance in video processing when shooting conditions are complex such as dynamic scenes and low light. Although existing algorithms have already achieved remarkable denoising performance, the inference time of them is usually impractical for real-time applications. In this paper, we propose a convolutional neural network architecture for video denoising. In contrast to other existing CNN-based methods, our approach utilizes different proportion convolutional kernel numbers in a block for extracting enriched features. Channel attention mechanism is integrated in the network to enhance the denoising performance. The network only needs three contiguous frames and noise map as inputs, which leads to a similar excellent running time to the state-of-the-art. We compare our method with different conventional algorithms VBM4D, VNLB and the state-of-the-art CNN-based method FastDVDnet. Experiment results indicate that our method outputs more convincing results in visual and more robustness than others in both peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) indexes.","PeriodicalId":127075,"journal":{"name":"2020 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS)","volume":"241 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Learning Enriched Features for Video Denoising with Convolutional Neural Network\",\"authors\":\"Xianfeng Tang, Peining Zhen, M. Kang, Hang Yi, Wei Wang, Hai-Bao Chen\",\"doi\":\"10.1109/APCCAS50809.2020.9301660\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Video denoising is of great significance in video processing when shooting conditions are complex such as dynamic scenes and low light. Although existing algorithms have already achieved remarkable denoising performance, the inference time of them is usually impractical for real-time applications. In this paper, we propose a convolutional neural network architecture for video denoising. In contrast to other existing CNN-based methods, our approach utilizes different proportion convolutional kernel numbers in a block for extracting enriched features. Channel attention mechanism is integrated in the network to enhance the denoising performance. The network only needs three contiguous frames and noise map as inputs, which leads to a similar excellent running time to the state-of-the-art. We compare our method with different conventional algorithms VBM4D, VNLB and the state-of-the-art CNN-based method FastDVDnet. Experiment results indicate that our method outputs more convincing results in visual and more robustness than others in both peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) indexes.\",\"PeriodicalId\":127075,\"journal\":{\"name\":\"2020 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS)\",\"volume\":\"241 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-12-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/APCCAS50809.2020.9301660\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/APCCAS50809.2020.9301660","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Learning Enriched Features for Video Denoising with Convolutional Neural Network
Video denoising is of great significance in video processing when shooting conditions are complex such as dynamic scenes and low light. Although existing algorithms have already achieved remarkable denoising performance, the inference time of them is usually impractical for real-time applications. In this paper, we propose a convolutional neural network architecture for video denoising. In contrast to other existing CNN-based methods, our approach utilizes different proportion convolutional kernel numbers in a block for extracting enriched features. Channel attention mechanism is integrated in the network to enhance the denoising performance. The network only needs three contiguous frames and noise map as inputs, which leads to a similar excellent running time to the state-of-the-art. We compare our method with different conventional algorithms VBM4D, VNLB and the state-of-the-art CNN-based method FastDVDnet. Experiment results indicate that our method outputs more convincing results in visual and more robustness than others in both peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) indexes.