{"title":"OSHFNet:用于土地利用分类的光学和SAR影像异构双分支动态融合网络","authors":"Chenfang Liu, Yuli Sun, Xianghui Zhang, Yanjie Xu, Lin Lei, Gangyao Kuang","doi":"10.1016/j.jag.2025.104609","DOIUrl":null,"url":null,"abstract":"<div><div>Optical and synthetic aperture radar (SAR) images are two of the most widely used remote sensing data sources, providing complementary but structurally consistent information. This complementarity has inspired significant research on their fusion. However, due to the huge difference in image representation between optical and SAR data, this difference will lead to inaccurate information expression when using the same structure to extract features, resulting in poor performance in classification tasks. Therefore, in the feature extraction stage, we analyze the respective advantageous features of optical and SAR images and propose a heterogeneous dual-branch network framework. Our framework exploits the rich local features of optical images and the global structural features of SAR images by using CNN and VMamba as their respective feature extractors. This heterogeneous feature extraction strategy effectively captures the complementary features of different modalities and provides a solid foundation for subsequent feature fusion. Second, in the feature fusion stage, we introduce a global-local dynamic gating fusion module. The use of multi-scale feature extraction and self-attention mechanism ensures comprehensive feature capture, while the dynamic gating mechanism enhances the integration of cross-modal complementary information. Finally, our method achieves excellent performance on medium and high-resolution datasets, showing robustness and adaptability at different resolutions. Notably, it significantly improves the overall classification accuracy while maintaining the accuracy of individual categories. For challenging categories such as roads, our method achieves significant improvements of about 15%.</div></div>","PeriodicalId":73423,"journal":{"name":"International journal of applied earth observation and geoinformation : ITC journal","volume":"141 ","pages":"Article 104609"},"PeriodicalIF":8.6000,"publicationDate":"2025-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"OSHFNet: A heterogeneous dual-branch dynamic fusion network of optical and SAR images for land use classification\",\"authors\":\"Chenfang Liu, Yuli Sun, Xianghui Zhang, Yanjie Xu, Lin Lei, Gangyao Kuang\",\"doi\":\"10.1016/j.jag.2025.104609\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Optical and synthetic aperture radar (SAR) images are two of the most widely used remote sensing data sources, providing complementary but structurally consistent information. This complementarity has inspired significant research on their fusion. However, due to the huge difference in image representation between optical and SAR data, this difference will lead to inaccurate information expression when using the same structure to extract features, resulting in poor performance in classification tasks. Therefore, in the feature extraction stage, we analyze the respective advantageous features of optical and SAR images and propose a heterogeneous dual-branch network framework. Our framework exploits the rich local features of optical images and the global structural features of SAR images by using CNN and VMamba as their respective feature extractors. This heterogeneous feature extraction strategy effectively captures the complementary features of different modalities and provides a solid foundation for subsequent feature fusion. Second, in the feature fusion stage, we introduce a global-local dynamic gating fusion module. The use of multi-scale feature extraction and self-attention mechanism ensures comprehensive feature capture, while the dynamic gating mechanism enhances the integration of cross-modal complementary information. Finally, our method achieves excellent performance on medium and high-resolution datasets, showing robustness and adaptability at different resolutions. Notably, it significantly improves the overall classification accuracy while maintaining the accuracy of individual categories. For challenging categories such as roads, our method achieves significant improvements of about 15%.</div></div>\",\"PeriodicalId\":73423,\"journal\":{\"name\":\"International journal of applied earth observation and geoinformation : ITC journal\",\"volume\":\"141 \",\"pages\":\"Article 104609\"},\"PeriodicalIF\":8.6000,\"publicationDate\":\"2025-06-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International journal of applied earth observation and geoinformation : ITC journal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1569843225002560\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"REMOTE SENSING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International journal of applied earth observation and geoinformation : ITC journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1569843225002560","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"REMOTE SENSING","Score":null,"Total":0}
OSHFNet: A heterogeneous dual-branch dynamic fusion network of optical and SAR images for land use classification
Optical and synthetic aperture radar (SAR) images are two of the most widely used remote sensing data sources, providing complementary but structurally consistent information. This complementarity has inspired significant research on their fusion. However, due to the huge difference in image representation between optical and SAR data, this difference will lead to inaccurate information expression when using the same structure to extract features, resulting in poor performance in classification tasks. Therefore, in the feature extraction stage, we analyze the respective advantageous features of optical and SAR images and propose a heterogeneous dual-branch network framework. Our framework exploits the rich local features of optical images and the global structural features of SAR images by using CNN and VMamba as their respective feature extractors. This heterogeneous feature extraction strategy effectively captures the complementary features of different modalities and provides a solid foundation for subsequent feature fusion. Second, in the feature fusion stage, we introduce a global-local dynamic gating fusion module. The use of multi-scale feature extraction and self-attention mechanism ensures comprehensive feature capture, while the dynamic gating mechanism enhances the integration of cross-modal complementary information. Finally, our method achieves excellent performance on medium and high-resolution datasets, showing robustness and adaptability at different resolutions. Notably, it significantly improves the overall classification accuracy while maintaining the accuracy of individual categories. For challenging categories such as roads, our method achieves significant improvements of about 15%.
期刊介绍:
The International Journal of Applied Earth Observation and Geoinformation publishes original papers that utilize earth observation data for natural resource and environmental inventory and management. These data primarily originate from remote sensing platforms, including satellites and aircraft, supplemented by surface and subsurface measurements. Addressing natural resources such as forests, agricultural land, soils, and water, as well as environmental concerns like biodiversity, land degradation, and hazards, the journal explores conceptual and data-driven approaches. It covers geoinformation themes like capturing, databasing, visualization, interpretation, data quality, and spatial uncertainty.