PromptMID: Modal invariant descriptors based on diffusion and vision foundation models for optical-SAR image matching

IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL
Han Nie , Bin Luo , Jun Liu , Zhitao Fu , Huan Zhou , Shuo Zhang , Weixing Liu
{"title":"PromptMID: Modal invariant descriptors based on diffusion and vision foundation models for optical-SAR image matching","authors":"Han Nie ,&nbsp;Bin Luo ,&nbsp;Jun Liu ,&nbsp;Zhitao Fu ,&nbsp;Huan Zhou ,&nbsp;Shuo Zhang ,&nbsp;Weixing Liu","doi":"10.1016/j.isprsjprs.2025.08.030","DOIUrl":null,"url":null,"abstract":"<div><div>The ideal goal of generalizable image matching is to achieve stable and efficient performance in unseen domains. However, many existing learning-based optical-SAR image matching methods, despite demonstrating effectiveness in specific scenarios, often exhibit limited generalization and face challenges in adapting to practical applications. Repeatedly training or fine-tuning matching models to address domain differences not only lacks elegance but also incurs additional computational overhead and data production costs. In recent years, foundation models have shown significant potential for enhancing generalization. However, the disparity in visual domains between natural and remote sensing images poses challenges for their direct application. Consequently, effectively leveraging foundation models to improve the generalization of optical-SAR image matching remains a critical challenge. To address these challenges, we propose PromptMID, a novel approach that constructs modality invariant descriptors using text prompts based on land use classification as priors information for optical and SAR image matching. PromptMID consists of several key stages. Firstly, we fine-tune the diffusion model (DM) using we collected optical images, SAR images, and text prompts data to obtain the PromptDM model. Secondly, we construct modality-invariant descriptors by integrating multi-scale latent diffusion features extracted from the fine-tuned PromptDM model with multi-scale features derived from pre-trained visual foundation models (VFMs). To efficiently fuse local–global and texture-semantic features of varying granularities, we design a feature aggregation module (FAM) that ensures comprehensive feature representation. Finally, the discriminative power of the descriptors is enhanced through contrastive learning loss functions, aiming to improve the robustness and generalization of matching. Extensive experiments conducted on optical-SAR image datasets from five diverse regions demonstrate that PromptMID outperforms state-of-the-art matching methods, achieving superior performance in both seen and unseen domains while exhibiting strong cross-domain generalization capabilities. The source code will be made publicly available <span><span>https://github.com/HanNieWHU/PromptMID</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"230 ","pages":"Pages 192-207"},"PeriodicalIF":12.2000,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003478","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0

Abstract

The ideal goal of generalizable image matching is to achieve stable and efficient performance in unseen domains. However, many existing learning-based optical-SAR image matching methods, despite demonstrating effectiveness in specific scenarios, often exhibit limited generalization and face challenges in adapting to practical applications. Repeatedly training or fine-tuning matching models to address domain differences not only lacks elegance but also incurs additional computational overhead and data production costs. In recent years, foundation models have shown significant potential for enhancing generalization. However, the disparity in visual domains between natural and remote sensing images poses challenges for their direct application. Consequently, effectively leveraging foundation models to improve the generalization of optical-SAR image matching remains a critical challenge. To address these challenges, we propose PromptMID, a novel approach that constructs modality invariant descriptors using text prompts based on land use classification as priors information for optical and SAR image matching. PromptMID consists of several key stages. Firstly, we fine-tune the diffusion model (DM) using we collected optical images, SAR images, and text prompts data to obtain the PromptDM model. Secondly, we construct modality-invariant descriptors by integrating multi-scale latent diffusion features extracted from the fine-tuned PromptDM model with multi-scale features derived from pre-trained visual foundation models (VFMs). To efficiently fuse local–global and texture-semantic features of varying granularities, we design a feature aggregation module (FAM) that ensures comprehensive feature representation. Finally, the discriminative power of the descriptors is enhanced through contrastive learning loss functions, aiming to improve the robustness and generalization of matching. Extensive experiments conducted on optical-SAR image datasets from five diverse regions demonstrate that PromptMID outperforms state-of-the-art matching methods, achieving superior performance in both seen and unseen domains while exhibiting strong cross-domain generalization capabilities. The source code will be made publicly available https://github.com/HanNieWHU/PromptMID.
基于扩散和视觉基础模型的模态不变描述子的光学sar图像匹配
广义图像匹配的理想目标是在不可见域实现稳定高效的性能。然而,许多现有的基于学习的光学sar图像匹配方法,尽管在特定场景下表现出有效性,但往往泛化有限,在适应实际应用方面面临挑战。反复训练或微调匹配模型来解决领域差异不仅缺乏优雅,而且还会产生额外的计算开销和数据生产成本。近年来,基础模型在增强泛化方面显示出巨大的潜力。然而,自然影像与遥感影像在视觉域上的差异给其直接应用带来了挑战。因此,有效地利用基础模型来提高光学- sar图像匹配的泛化仍然是一个关键的挑战。为了解决这些挑战,我们提出了一种新的方法PromptMID,它使用基于土地利用分类的文本提示作为光学和SAR图像匹配的先验信息来构建模态不变描述符。PromptMID由几个关键阶段组成。首先,我们利用收集的光学图像、SAR图像和文本提示数据对扩散模型(DM)进行微调,得到PromptDM模型。其次,我们通过整合从微调后的PromptDM模型中提取的多尺度潜在扩散特征和从预训练的视觉基础模型(VFMs)中提取的多尺度特征来构建模态不变描述子。为了有效地融合局部全局特征和不同粒度的纹理语义特征,我们设计了一个特征聚合模块(FAM)来保证特征的全面表达。最后,通过对比学习损失函数增强描述符的判别能力,提高匹配的鲁棒性和泛化性。对来自五个不同地区的光学sar图像数据集进行的大量实验表明,PromptMID优于最先进的匹配方法,在可见域和不可见域都取得了卓越的性能,同时表现出强大的跨域泛化能力。源代码将公开提供https://github.com/HanNieWHU/PromptMID。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
ISPRS Journal of Photogrammetry and Remote Sensing
ISPRS Journal of Photogrammetry and Remote Sensing 工程技术-成像科学与照相技术
CiteScore
21.00
自引率
6.30%
发文量
273
审稿时长
40 days
期刊介绍: The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive. P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields. In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信