{"title":"STARNet: Low-light video enhancement using spatio-temporal consistency aggregation","authors":"Zhe Wu , Zehua Sheng , Xue Zhang , Si-Yuan Cao , Runmin Zhang , Beinan Yu , Chenghao Zhang , Bailin Yang , Hui-Liang Shen","doi":"10.1016/j.patcog.2024.111180","DOIUrl":null,"url":null,"abstract":"<div><div>In low-light environments, capturing high-quality videos is an imaging challenge due to the limited number of photons. Previous low-light enhancement approaches usually result in over-smoothed details, temporal flickers, and color deviation. We propose STARNet, an end-to-end video enhancement network that leverages temporal consistency aggregation to address these issues. We introduce a spatio-temporal consistency aggregator, which extracts structures from multiple frames in hidden space to overcome detail corruption and temporal flickers. It parameterizes neighboring frames to extract and align consistent features, and then selectively fuses consistent features to restore clear structures. To further enhance temporal consistency, we develop a local temporal consistency constraint with robustness against the warping error from motion estimation. Furthermore, we employ a normalized low-frequency color constraint to regularize the color as the normal-light condition. Extensive experimental results on real datasets show that the proposed method achieves better detail fidelity, color accuracy, and temporal consistency, outperforming state-of-the-art approaches.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"160 ","pages":"Article 111180"},"PeriodicalIF":7.5000,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0031320324009312","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
In low-light environments, capturing high-quality videos is an imaging challenge due to the limited number of photons. Previous low-light enhancement approaches usually result in over-smoothed details, temporal flickers, and color deviation. We propose STARNet, an end-to-end video enhancement network that leverages temporal consistency aggregation to address these issues. We introduce a spatio-temporal consistency aggregator, which extracts structures from multiple frames in hidden space to overcome detail corruption and temporal flickers. It parameterizes neighboring frames to extract and align consistent features, and then selectively fuses consistent features to restore clear structures. To further enhance temporal consistency, we develop a local temporal consistency constraint with robustness against the warping error from motion estimation. Furthermore, we employ a normalized low-frequency color constraint to regularize the color as the normal-light condition. Extensive experimental results on real datasets show that the proposed method achieves better detail fidelity, color accuracy, and temporal consistency, outperforming state-of-the-art approaches.
期刊介绍:
The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.