{"title":"PAMSNet: A point annotation-driven multi-source network for remote sensing semantic segmentation","authors":"Yuanhao Zhao , Mingming Jia , Genyun Sun , Aizhu Zhang","doi":"10.1016/j.isprsjprs.2025.07.035","DOIUrl":null,"url":null,"abstract":"<div><div>Multi-source data semantic segmentation has proven to be an effective means of improving classification accuracy in remote sensing. With the rapid development of deep learning, the demand for large amounts of high-quality labeled samples has become a major bottleneck, limiting the broader application of these techniques. Weakly supervised learning has attracted increasing attention by reducing annotation costs. However, existing weakly supervised methods often suffer from limited accuracy. Effectively exploiting complementary information from multi-source remote sensing data using only a small number of labeled points remains a significant challenge. In this paper, we propose a novel architecture, named Point Annotation- Driven Multi-source Segmentation Network (PAMSNet), which leverages point annotations to effectively capture and integrate complementary features from multi-source remote sensing data. PAMSNet includes a Multi-source Feature Encoder and a Cross-Resolution Feature Integration (CRFI) module. The Multi-source Feature Encoder captures complementary global and local features using lightweight convolutional Global-Local Multi-source (GLMS) modules. And the boundary and spectral detail robustness are improved through Spectral-Edge Enhancement (SEE) modules, which effectively mitigate the impact of noise on segmentation accuracy. The CRFI module replaces conventional decoding structures by combining convolutional and Transformer mechanisms, enabling efficient cross-scale feature integration and improving the ability to identify multi-scale objects with reduced computational demands. Extensive experiments on the Vaihingen, WHU-IS, and WHU-OPT-SAR datasets validate the effectiveness of PAMSNet for weakly supervised multi-source segmentation as well as the validity of the proposed module. PAMSNet achieves state-of-the-art performance, with MIoU improvements of 2.4%, 2.1%, and 3.16% on three datasets, using only 0.01% point annotations. Additionally, PAMSNet can effectively balance the performance as well as the operational efficiency of the model compared to existing methods, which further promotes the application of deep learning in remote sensing image mapping.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"229 ","pages":"Pages 1-16"},"PeriodicalIF":12.2000,"publicationDate":"2025-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003041","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Multi-source data semantic segmentation has proven to be an effective means of improving classification accuracy in remote sensing. With the rapid development of deep learning, the demand for large amounts of high-quality labeled samples has become a major bottleneck, limiting the broader application of these techniques. Weakly supervised learning has attracted increasing attention by reducing annotation costs. However, existing weakly supervised methods often suffer from limited accuracy. Effectively exploiting complementary information from multi-source remote sensing data using only a small number of labeled points remains a significant challenge. In this paper, we propose a novel architecture, named Point Annotation- Driven Multi-source Segmentation Network (PAMSNet), which leverages point annotations to effectively capture and integrate complementary features from multi-source remote sensing data. PAMSNet includes a Multi-source Feature Encoder and a Cross-Resolution Feature Integration (CRFI) module. The Multi-source Feature Encoder captures complementary global and local features using lightweight convolutional Global-Local Multi-source (GLMS) modules. And the boundary and spectral detail robustness are improved through Spectral-Edge Enhancement (SEE) modules, which effectively mitigate the impact of noise on segmentation accuracy. The CRFI module replaces conventional decoding structures by combining convolutional and Transformer mechanisms, enabling efficient cross-scale feature integration and improving the ability to identify multi-scale objects with reduced computational demands. Extensive experiments on the Vaihingen, WHU-IS, and WHU-OPT-SAR datasets validate the effectiveness of PAMSNet for weakly supervised multi-source segmentation as well as the validity of the proposed module. PAMSNet achieves state-of-the-art performance, with MIoU improvements of 2.4%, 2.1%, and 3.16% on three datasets, using only 0.01% point annotations. Additionally, PAMSNet can effectively balance the performance as well as the operational efficiency of the model compared to existing methods, which further promotes the application of deep learning in remote sensing image mapping.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.