Kui Jiang;Qiong Wang;Zhaoyi An;Zheng Wang;Cong Zhang;Chia-Wen Lin
{"title":"Mutual Retinex:结合变换器和 CNN 增强图像效果","authors":"Kui Jiang;Qiong Wang;Zhaoyi An;Zheng Wang;Cong Zhang;Chia-Wen Lin","doi":"10.1109/TETCI.2024.3369321","DOIUrl":null,"url":null,"abstract":"Images captured in low-light or underwater environments are often accompanied by significant degradation, which can negatively impact the quality and performance of downstream tasks. While convolutional neural networks (CNNs) and Transformer architectures have made significant progress in computer vision tasks, there are few efforts to harmonize them into a more concise framework for enhancing such images. To this end, this study proposes to aggregate the individual capability of self-attention (SA) and CNNs for accurate perturbation removal while preserving background contents. Based on this, we carry forward a Retinex-based framework, dubbed as Mutual Retinex, where a two-branch structure is designed to characterize the specific knowledge of reflectance and illumination components while removing the perturbation. To maximize its potential, Mutual Retinex is equipped with a new mutual learning mechanism, involving an elaborately designed mutual representation module (MRM). In MRM, the complementary information between reflectance and illumination components are encoded and used to refine each other. Through the complementary learning via the mutual representation, the enhanced results generated by our model exhibit superior color consistency and naturalness. Extensive experiments have shown the significant superiority of our mutual learning based method over thirteen competitors on the low-light task and ten methods on the underwater image enhancement task. In particular, our proposed Mutual Retinex respectively surpasses the state-of-the-art method MIRNet-v2 by 0.90 dB and 2.46 dB in PSNR on the LOL 1000 and FIVEK datasets, while with only 19.8% model parameters.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"8 3","pages":"2240-2252"},"PeriodicalIF":5.3000,"publicationDate":"2024-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Mutual Retinex: Combining Transformer and CNN for Image Enhancement\",\"authors\":\"Kui Jiang;Qiong Wang;Zhaoyi An;Zheng Wang;Cong Zhang;Chia-Wen Lin\",\"doi\":\"10.1109/TETCI.2024.3369321\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Images captured in low-light or underwater environments are often accompanied by significant degradation, which can negatively impact the quality and performance of downstream tasks. While convolutional neural networks (CNNs) and Transformer architectures have made significant progress in computer vision tasks, there are few efforts to harmonize them into a more concise framework for enhancing such images. To this end, this study proposes to aggregate the individual capability of self-attention (SA) and CNNs for accurate perturbation removal while preserving background contents. Based on this, we carry forward a Retinex-based framework, dubbed as Mutual Retinex, where a two-branch structure is designed to characterize the specific knowledge of reflectance and illumination components while removing the perturbation. To maximize its potential, Mutual Retinex is equipped with a new mutual learning mechanism, involving an elaborately designed mutual representation module (MRM). In MRM, the complementary information between reflectance and illumination components are encoded and used to refine each other. Through the complementary learning via the mutual representation, the enhanced results generated by our model exhibit superior color consistency and naturalness. Extensive experiments have shown the significant superiority of our mutual learning based method over thirteen competitors on the low-light task and ten methods on the underwater image enhancement task. In particular, our proposed Mutual Retinex respectively surpasses the state-of-the-art method MIRNet-v2 by 0.90 dB and 2.46 dB in PSNR on the LOL 1000 and FIVEK datasets, while with only 19.8% model parameters.\",\"PeriodicalId\":13135,\"journal\":{\"name\":\"IEEE Transactions on Emerging Topics in Computational Intelligence\",\"volume\":\"8 3\",\"pages\":\"2240-2252\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2024-03-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Emerging Topics in Computational Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10462575/\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10462575/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Mutual Retinex: Combining Transformer and CNN for Image Enhancement
Images captured in low-light or underwater environments are often accompanied by significant degradation, which can negatively impact the quality and performance of downstream tasks. While convolutional neural networks (CNNs) and Transformer architectures have made significant progress in computer vision tasks, there are few efforts to harmonize them into a more concise framework for enhancing such images. To this end, this study proposes to aggregate the individual capability of self-attention (SA) and CNNs for accurate perturbation removal while preserving background contents. Based on this, we carry forward a Retinex-based framework, dubbed as Mutual Retinex, where a two-branch structure is designed to characterize the specific knowledge of reflectance and illumination components while removing the perturbation. To maximize its potential, Mutual Retinex is equipped with a new mutual learning mechanism, involving an elaborately designed mutual representation module (MRM). In MRM, the complementary information between reflectance and illumination components are encoded and used to refine each other. Through the complementary learning via the mutual representation, the enhanced results generated by our model exhibit superior color consistency and naturalness. Extensive experiments have shown the significant superiority of our mutual learning based method over thirteen competitors on the low-light task and ten methods on the underwater image enhancement task. In particular, our proposed Mutual Retinex respectively surpasses the state-of-the-art method MIRNet-v2 by 0.90 dB and 2.46 dB in PSNR on the LOL 1000 and FIVEK datasets, while with only 19.8% model parameters.
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.