Zhiling Geng , Haibo Liu , Puhong Duan , Xiaohui Wei , Shutao Li
{"title":"Feature-based multimodal remote sensing image matching: Benchmark and state-of-the-art","authors":"Zhiling Geng , Haibo Liu , Puhong Duan , Xiaohui Wei , Shutao Li","doi":"10.1016/j.isprsjprs.2025.08.028","DOIUrl":null,"url":null,"abstract":"<div><div>Multimodal remote sensing image matching (MRSIM) is a crucial prerequisite in the remote sensing field, aiming to align images captured by different sensors to facilitate subsequent interpretation and analysis. In recent years, numerous efforts have been made to achieve feature-based MRSIM. However, there is a lack of a comprehensive review of advanced feature-based MRSIM methods and a comparison of their performance on diverse datasets. Additionally, existing datasets often have some limitations in terms of modality diversity and ground truth completeness, which prevent the validation of the performance of algorithms. This paper first provides an extensive overview of latest advances based on the general framework of feature-based MRSIM methods. Then, we summarize existing MRSIM datasets, and construct the HNU-DATASET, including four types of common MRSIM pairs and ground-truth annotations of each image pair. Finally, to ensure a comprehensive evaluation, several representative open-source methods, such as radiation-variation insensitive feature transform (RIFT) and histogram of absolute phase consistency gradients (HAPCG), are employed to benchmark performance on both the proposed HNU-DATASET and multiple publicly available datasets. The experimental results can serve as a valuable reference for future research, which can promote the development of advanced multimodal remote sensing.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"229 ","pages":"Pages 285-302"},"PeriodicalIF":12.2000,"publicationDate":"2025-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003399","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Multimodal remote sensing image matching (MRSIM) is a crucial prerequisite in the remote sensing field, aiming to align images captured by different sensors to facilitate subsequent interpretation and analysis. In recent years, numerous efforts have been made to achieve feature-based MRSIM. However, there is a lack of a comprehensive review of advanced feature-based MRSIM methods and a comparison of their performance on diverse datasets. Additionally, existing datasets often have some limitations in terms of modality diversity and ground truth completeness, which prevent the validation of the performance of algorithms. This paper first provides an extensive overview of latest advances based on the general framework of feature-based MRSIM methods. Then, we summarize existing MRSIM datasets, and construct the HNU-DATASET, including four types of common MRSIM pairs and ground-truth annotations of each image pair. Finally, to ensure a comprehensive evaluation, several representative open-source methods, such as radiation-variation insensitive feature transform (RIFT) and histogram of absolute phase consistency gradients (HAPCG), are employed to benchmark performance on both the proposed HNU-DATASET and multiple publicly available datasets. The experimental results can serve as a valuable reference for future research, which can promote the development of advanced multimodal remote sensing.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.