{"title":"MSFFNet: Multi-stream feature fusion network for underwater image enhancement","authors":"Peng Lin, Zihao Fan, Yafei Wang, Xudong Sun, Yuán-Ruì Yáng, Xianping Fu","doi":"10.1016/j.displa.2025.103023","DOIUrl":null,"url":null,"abstract":"<div><div>Deep learning-based image processing methods have achieved remarkable success in improving the quality of underwater images. These methods usually extract features from different receptive fields through downsampling operations, and then enhance underwater images through upsampling operations. However, these operations of downsampling and upsampling inevitably disrupt the relations of neighboring pixels in raw underwater images, leading to the loss of image details. Given this, a multi-stream feature fusion network, dubbed MSFFNet, is proposed to enrich details, correct colors, and enhance contrast of degraded underwater images. In MSFFNet, the multi-stream feature estimation block is carefully constructed, which separately takes original resolution feature maps and low-resolution feature maps as inputs. The multi-stream feature estimation block proficiently preserves the details information of the original underwater image while extracting high-level features. Besides, a coordinate residual block is designed to emphasize valuable features and suppress noises based on position knowledge. A local–global feature fusion block is presented for selectively fusing the complementary multi-scale features. Finally, extensive comparative experiments on real underwater images and synthetic underwater images demonstrate that the proposed MSFFNet has superior performance on underwater image enhancement tasks.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103023"},"PeriodicalIF":3.7000,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Displays","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0141938225000605","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
MSFFNet: Multi-stream feature fusion network for underwater image enhancement
Deep learning-based image processing methods have achieved remarkable success in improving the quality of underwater images. These methods usually extract features from different receptive fields through downsampling operations, and then enhance underwater images through upsampling operations. However, these operations of downsampling and upsampling inevitably disrupt the relations of neighboring pixels in raw underwater images, leading to the loss of image details. Given this, a multi-stream feature fusion network, dubbed MSFFNet, is proposed to enrich details, correct colors, and enhance contrast of degraded underwater images. In MSFFNet, the multi-stream feature estimation block is carefully constructed, which separately takes original resolution feature maps and low-resolution feature maps as inputs. The multi-stream feature estimation block proficiently preserves the details information of the original underwater image while extracting high-level features. Besides, a coordinate residual block is designed to emphasize valuable features and suppress noises based on position knowledge. A local–global feature fusion block is presented for selectively fusing the complementary multi-scale features. Finally, extensive comparative experiments on real underwater images and synthetic underwater images demonstrate that the proposed MSFFNet has superior performance on underwater image enhancement tasks.
期刊介绍:
Displays is the international journal covering the research and development of display technology, its effective presentation and perception of information, and applications and systems including display-human interface.
Technical papers on practical developments in Displays technology provide an effective channel to promote greater understanding and cross-fertilization across the diverse disciplines of the Displays community. Original research papers solving ergonomics issues at the display-human interface advance effective presentation of information. Tutorial papers covering fundamentals intended for display technologies and human factor engineers new to the field will also occasionally featured.