{"title":"Saliency supervised masked autoencoder pretrained salient location mining network for remote sensing image salient object detection","authors":"Yuxiang Fu , Wei Fang , Victor S. Sheng","doi":"10.1016/j.isprsjprs.2025.03.025","DOIUrl":null,"url":null,"abstract":"<div><div>Remote sensing image salient object detection (RSI-SOD), as an emerging topic in computer vision, has significant applications across various sectors, such as urban planning, environmental monitoring and disaster management, etc. In recent years, RSI-SOD has seen significant advancements, largely due to advanced representation learning methods and better architectures, such as convolutional neural networks and vision transformers. While current methods predominantly rely on supervised learning, there is potential for enhancement through self-supervised learning approaches, like masked autoencoder. However, we observed that the conventional use of masked autoencoder for pretraining encoders through masked image reconstruction yields subpar results in the context of RSI-SOD. To this end, we propose a novel approach: saliency supervised masked autoencoder (SSMAE) and a corresponding salient location mining network (SLMNet), which is pretrained by SSMAE for the task of RSI-SOD. SSMAE first uses masked autoencoder to reconstruct the masked image, and then employs SLMNet to predict saliency map from the reconstructed image, where saliency supervision is adopted to enable SLMNet to learn robust saliency prior knowledge. SLMNet has three major components: encoder, salient location mining module (SLMM) and the decoder. Specifically, SLMM employs residual multi-level fusion structure to mine the locations of salient objects from multi-scale features produced by the encoder. Later, the decoder fuses the multi-level features from SLMM and encoder to generate the prediction results. Comprehensive experiments on three public datasets demonstrate that our proposed method surpasses the state-of-the-art methods. Code is available at: <span><span>https://github.com/Voruarn/SLMNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"224 ","pages":"Pages 222-234"},"PeriodicalIF":10.6000,"publicationDate":"2025-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625001236","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Remote sensing image salient object detection (RSI-SOD), as an emerging topic in computer vision, has significant applications across various sectors, such as urban planning, environmental monitoring and disaster management, etc. In recent years, RSI-SOD has seen significant advancements, largely due to advanced representation learning methods and better architectures, such as convolutional neural networks and vision transformers. While current methods predominantly rely on supervised learning, there is potential for enhancement through self-supervised learning approaches, like masked autoencoder. However, we observed that the conventional use of masked autoencoder for pretraining encoders through masked image reconstruction yields subpar results in the context of RSI-SOD. To this end, we propose a novel approach: saliency supervised masked autoencoder (SSMAE) and a corresponding salient location mining network (SLMNet), which is pretrained by SSMAE for the task of RSI-SOD. SSMAE first uses masked autoencoder to reconstruct the masked image, and then employs SLMNet to predict saliency map from the reconstructed image, where saliency supervision is adopted to enable SLMNet to learn robust saliency prior knowledge. SLMNet has three major components: encoder, salient location mining module (SLMM) and the decoder. Specifically, SLMM employs residual multi-level fusion structure to mine the locations of salient objects from multi-scale features produced by the encoder. Later, the decoder fuses the multi-level features from SLMM and encoder to generate the prediction results. Comprehensive experiments on three public datasets demonstrate that our proposed method surpasses the state-of-the-art methods. Code is available at: https://github.com/Voruarn/SLMNet.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.