Wenjie Sun , Yujie Lei , Danfeng Hong , Zhongwen Hu , Qingquan Li , Jie Zhang
{"title":"RSProtoSemiSeg:基于概率分布原型的高空间分辨率遥感图像半监督语义分割","authors":"Wenjie Sun , Yujie Lei , Danfeng Hong , Zhongwen Hu , Qingquan Li , Jie Zhang","doi":"10.1016/j.isprsjprs.2025.07.040","DOIUrl":null,"url":null,"abstract":"<div><div>Semi-supervised semantic segmentation of high spatial resolution remote sensing images aims to mitigate the reliance on labeled data by using limited labeled data alongside extensive unlabeled data. This approach significantly reduces the dependency on labeled images and annotation costs and mitigates the challenge of obtaining large-scale labeled datasets. Current semi-supervised methods for remote sensing images semantic segmentation mainly focus on improving pseudo-label quality. However, the inherent noise in pseudo-labels remains a critical issue, leading to persistent inaccuracies. Furthermore, the high intra-class variance in such image complicate pixel-wise label propagation, exacerbating pseudo-label errors and substantially constraining segmentation accuracy. To tackle these limitations, we propose a contrastive learning framework based on a mixture of Gaussian mixture distributions. Our approach uses a mixture probability distribution prototype predictor to adaptively regulate the influence of intra-class prototypes on pixel representations, mapping features to a multivariate Gaussian model to mitigate pseudo-label inaccuracies. We also introduce a novel mixture contrastive loss function to guide pseudo-labeled pixels towards intra-class prototypes and away from inter-class prototypes while repelling, thereby enhancing representation fidelity. Extensive experiments on three remote sensing semantic segmentation datasets demonstrate the efficacy of our approach. Compared to baseline models, our method achieves mIoU improvements ranging from 0.26% to 1.86%. Compared to the state-of-the-art DWL model, it achieves even higher mIoU improvements, ranging from 0.37% to 2.36%. These results confirm the effectiveness of our approach over existing semi-supervised segmentation methods. The code will be available at <span><span>https://github.com/aitointerp/rsprotosemiseg</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 771-784"},"PeriodicalIF":12.2000,"publicationDate":"2025-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"RSProtoSemiSeg: Semi-supervised semantic segmentation of high spatial resolution remote sensing images with probabilistic distribution prototypes\",\"authors\":\"Wenjie Sun , Yujie Lei , Danfeng Hong , Zhongwen Hu , Qingquan Li , Jie Zhang\",\"doi\":\"10.1016/j.isprsjprs.2025.07.040\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Semi-supervised semantic segmentation of high spatial resolution remote sensing images aims to mitigate the reliance on labeled data by using limited labeled data alongside extensive unlabeled data. This approach significantly reduces the dependency on labeled images and annotation costs and mitigates the challenge of obtaining large-scale labeled datasets. Current semi-supervised methods for remote sensing images semantic segmentation mainly focus on improving pseudo-label quality. However, the inherent noise in pseudo-labels remains a critical issue, leading to persistent inaccuracies. Furthermore, the high intra-class variance in such image complicate pixel-wise label propagation, exacerbating pseudo-label errors and substantially constraining segmentation accuracy. To tackle these limitations, we propose a contrastive learning framework based on a mixture of Gaussian mixture distributions. Our approach uses a mixture probability distribution prototype predictor to adaptively regulate the influence of intra-class prototypes on pixel representations, mapping features to a multivariate Gaussian model to mitigate pseudo-label inaccuracies. We also introduce a novel mixture contrastive loss function to guide pseudo-labeled pixels towards intra-class prototypes and away from inter-class prototypes while repelling, thereby enhancing representation fidelity. Extensive experiments on three remote sensing semantic segmentation datasets demonstrate the efficacy of our approach. Compared to baseline models, our method achieves mIoU improvements ranging from 0.26% to 1.86%. Compared to the state-of-the-art DWL model, it achieves even higher mIoU improvements, ranging from 0.37% to 2.36%. These results confirm the effectiveness of our approach over existing semi-supervised segmentation methods. The code will be available at <span><span>https://github.com/aitointerp/rsprotosemiseg</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50269,\"journal\":{\"name\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"volume\":\"228 \",\"pages\":\"Pages 771-784\"},\"PeriodicalIF\":12.2000,\"publicationDate\":\"2025-08-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0924271625003107\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"GEOGRAPHY, PHYSICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003107","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
RSProtoSemiSeg: Semi-supervised semantic segmentation of high spatial resolution remote sensing images with probabilistic distribution prototypes
Semi-supervised semantic segmentation of high spatial resolution remote sensing images aims to mitigate the reliance on labeled data by using limited labeled data alongside extensive unlabeled data. This approach significantly reduces the dependency on labeled images and annotation costs and mitigates the challenge of obtaining large-scale labeled datasets. Current semi-supervised methods for remote sensing images semantic segmentation mainly focus on improving pseudo-label quality. However, the inherent noise in pseudo-labels remains a critical issue, leading to persistent inaccuracies. Furthermore, the high intra-class variance in such image complicate pixel-wise label propagation, exacerbating pseudo-label errors and substantially constraining segmentation accuracy. To tackle these limitations, we propose a contrastive learning framework based on a mixture of Gaussian mixture distributions. Our approach uses a mixture probability distribution prototype predictor to adaptively regulate the influence of intra-class prototypes on pixel representations, mapping features to a multivariate Gaussian model to mitigate pseudo-label inaccuracies. We also introduce a novel mixture contrastive loss function to guide pseudo-labeled pixels towards intra-class prototypes and away from inter-class prototypes while repelling, thereby enhancing representation fidelity. Extensive experiments on three remote sensing semantic segmentation datasets demonstrate the efficacy of our approach. Compared to baseline models, our method achieves mIoU improvements ranging from 0.26% to 1.86%. Compared to the state-of-the-art DWL model, it achieves even higher mIoU improvements, ranging from 0.37% to 2.36%. These results confirm the effectiveness of our approach over existing semi-supervised segmentation methods. The code will be available at https://github.com/aitointerp/rsprotosemiseg.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.