Zheng Cong , Yifeng Zhou , Li Wu , Lin Tian , Zhipeng Chen , Minglei Guan , Li He
{"title":"PGF-Net:融合自关注物理成像模型的鲁棒水下特征检测","authors":"Zheng Cong , Yifeng Zhou , Li Wu , Lin Tian , Zhipeng Chen , Minglei Guan , Li He","doi":"10.1016/j.inffus.2025.103732","DOIUrl":null,"url":null,"abstract":"<div><div>Robust feature detection in underwater environments is severely impeded by image degradation from light absorption and scattering. Traditional algorithms fail in these low-contrast, blurred conditions, while deep learning methods suffer from the domain gap between terrestrial and underwater imagery and a scarcity of annotated data. To address these challenges, this paper introduces PGF-Net, a systematic framework that fuses physical imaging principles with deep learning. The framework leverages a dual-fusion strategy: First, a parametric underwater imaging model is proposed to guide the synthesis of a large-scale, physically realistic training dataset, effectively injecting prior knowledge of the degradation process into the data domain. Second, a novel detection network architecture is designed, which incorporates a self-attention mechanism to fuse local features with global contextual information, enhancing robustness against detail loss. This end-to-end network is trained on the synthesized data using a curriculum learning strategy, progressing from mild to severe degradation conditions. Extensive experiments on public datasets demonstrate that PGF-Net significantly outperforms classic and state-of-the-art deep learning methods in both keypoint detection and matching, particularly in turbid water. The proposed framework validates the efficacy of integrating physical priors with data-driven models for challenging computer vision tasks and provides a robust solution for underwater visual perception.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103732"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"PGF-Net: fusing physical imaging model with self-attention for robust underwater feature detection\",\"authors\":\"Zheng Cong , Yifeng Zhou , Li Wu , Lin Tian , Zhipeng Chen , Minglei Guan , Li He\",\"doi\":\"10.1016/j.inffus.2025.103732\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Robust feature detection in underwater environments is severely impeded by image degradation from light absorption and scattering. Traditional algorithms fail in these low-contrast, blurred conditions, while deep learning methods suffer from the domain gap between terrestrial and underwater imagery and a scarcity of annotated data. To address these challenges, this paper introduces PGF-Net, a systematic framework that fuses physical imaging principles with deep learning. The framework leverages a dual-fusion strategy: First, a parametric underwater imaging model is proposed to guide the synthesis of a large-scale, physically realistic training dataset, effectively injecting prior knowledge of the degradation process into the data domain. Second, a novel detection network architecture is designed, which incorporates a self-attention mechanism to fuse local features with global contextual information, enhancing robustness against detail loss. This end-to-end network is trained on the synthesized data using a curriculum learning strategy, progressing from mild to severe degradation conditions. Extensive experiments on public datasets demonstrate that PGF-Net significantly outperforms classic and state-of-the-art deep learning methods in both keypoint detection and matching, particularly in turbid water. The proposed framework validates the efficacy of integrating physical priors with data-driven models for challenging computer vision tasks and provides a robust solution for underwater visual perception.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"127 \",\"pages\":\"Article 103732\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-09-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525007948\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525007948","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
PGF-Net: fusing physical imaging model with self-attention for robust underwater feature detection
Robust feature detection in underwater environments is severely impeded by image degradation from light absorption and scattering. Traditional algorithms fail in these low-contrast, blurred conditions, while deep learning methods suffer from the domain gap between terrestrial and underwater imagery and a scarcity of annotated data. To address these challenges, this paper introduces PGF-Net, a systematic framework that fuses physical imaging principles with deep learning. The framework leverages a dual-fusion strategy: First, a parametric underwater imaging model is proposed to guide the synthesis of a large-scale, physically realistic training dataset, effectively injecting prior knowledge of the degradation process into the data domain. Second, a novel detection network architecture is designed, which incorporates a self-attention mechanism to fuse local features with global contextual information, enhancing robustness against detail loss. This end-to-end network is trained on the synthesized data using a curriculum learning strategy, progressing from mild to severe degradation conditions. Extensive experiments on public datasets demonstrate that PGF-Net significantly outperforms classic and state-of-the-art deep learning methods in both keypoint detection and matching, particularly in turbid water. The proposed framework validates the efficacy of integrating physical priors with data-driven models for challenging computer vision tasks and provides a robust solution for underwater visual perception.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.