Han Nie , Bin Luo , Jun Liu , Zhitao Fu , Huan Zhou , Shuo Zhang , Weixing Liu
{"title":"PromptMID: Modal invariant descriptors based on diffusion and vision foundation models for optical-SAR image matching","authors":"Han Nie , Bin Luo , Jun Liu , Zhitao Fu , Huan Zhou , Shuo Zhang , Weixing Liu","doi":"10.1016/j.isprsjprs.2025.08.030","DOIUrl":null,"url":null,"abstract":"<div><div>The ideal goal of generalizable image matching is to achieve stable and efficient performance in unseen domains. However, many existing learning-based optical-SAR image matching methods, despite demonstrating effectiveness in specific scenarios, often exhibit limited generalization and face challenges in adapting to practical applications. Repeatedly training or fine-tuning matching models to address domain differences not only lacks elegance but also incurs additional computational overhead and data production costs. In recent years, foundation models have shown significant potential for enhancing generalization. However, the disparity in visual domains between natural and remote sensing images poses challenges for their direct application. Consequently, effectively leveraging foundation models to improve the generalization of optical-SAR image matching remains a critical challenge. To address these challenges, we propose PromptMID, a novel approach that constructs modality invariant descriptors using text prompts based on land use classification as priors information for optical and SAR image matching. PromptMID consists of several key stages. Firstly, we fine-tune the diffusion model (DM) using we collected optical images, SAR images, and text prompts data to obtain the PromptDM model. Secondly, we construct modality-invariant descriptors by integrating multi-scale latent diffusion features extracted from the fine-tuned PromptDM model with multi-scale features derived from pre-trained visual foundation models (VFMs). To efficiently fuse local–global and texture-semantic features of varying granularities, we design a feature aggregation module (FAM) that ensures comprehensive feature representation. Finally, the discriminative power of the descriptors is enhanced through contrastive learning loss functions, aiming to improve the robustness and generalization of matching. Extensive experiments conducted on optical-SAR image datasets from five diverse regions demonstrate that PromptMID outperforms state-of-the-art matching methods, achieving superior performance in both seen and unseen domains while exhibiting strong cross-domain generalization capabilities. The source code will be made publicly available <span><span>https://github.com/HanNieWHU/PromptMID</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"230 ","pages":"Pages 192-207"},"PeriodicalIF":12.2000,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003478","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
The ideal goal of generalizable image matching is to achieve stable and efficient performance in unseen domains. However, many existing learning-based optical-SAR image matching methods, despite demonstrating effectiveness in specific scenarios, often exhibit limited generalization and face challenges in adapting to practical applications. Repeatedly training or fine-tuning matching models to address domain differences not only lacks elegance but also incurs additional computational overhead and data production costs. In recent years, foundation models have shown significant potential for enhancing generalization. However, the disparity in visual domains between natural and remote sensing images poses challenges for their direct application. Consequently, effectively leveraging foundation models to improve the generalization of optical-SAR image matching remains a critical challenge. To address these challenges, we propose PromptMID, a novel approach that constructs modality invariant descriptors using text prompts based on land use classification as priors information for optical and SAR image matching. PromptMID consists of several key stages. Firstly, we fine-tune the diffusion model (DM) using we collected optical images, SAR images, and text prompts data to obtain the PromptDM model. Secondly, we construct modality-invariant descriptors by integrating multi-scale latent diffusion features extracted from the fine-tuned PromptDM model with multi-scale features derived from pre-trained visual foundation models (VFMs). To efficiently fuse local–global and texture-semantic features of varying granularities, we design a feature aggregation module (FAM) that ensures comprehensive feature representation. Finally, the discriminative power of the descriptors is enhanced through contrastive learning loss functions, aiming to improve the robustness and generalization of matching. Extensive experiments conducted on optical-SAR image datasets from five diverse regions demonstrate that PromptMID outperforms state-of-the-art matching methods, achieving superior performance in both seen and unseen domains while exhibiting strong cross-domain generalization capabilities. The source code will be made publicly available https://github.com/HanNieWHU/PromptMID.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.