{"title":"Prompt-Based Granularity-Unified Representation Network for Remote Sensing Image-Text Matching","authors":"Minhan Hu;Keke Yang;Jing Li","doi":"10.1109/JSTARS.2025.3555639","DOIUrl":null,"url":null,"abstract":"Remote sensing (RS) image–text matching has gained significant attention for its promising potential. Despite great advancements, accurately matching RS images (RSIs) and captions remains challenging due to the significant multimodal gap and inherent characteristics of RS data. Many approaches use complex models to extract global features to handle semantic redundancy and varying scales in RSIs, but losing important details in RSIs and captions. While some methods align between fine-grained local features, but overlooking the semantic granularity differences between fine-grained features. Fine-grained features in RSIs typically capture only a small fraction of the overall semantics, whereas those in captions convey more comprehensive and abstract semantics. Therefore, we propose the prompt-based granularity-unified representation network, an end-to-end framework designed to mitigate the multimodal semantic granularity difference and achieve comprehensive alignment. Our approach includes two key modules: 1) the prompt-based feature aggregator, which dynamically aggregates fine-grained features into several granularity-unified tokens with fully semantic, and 2) the text-guided vision modulation, which further enhances visual representations by modulating the visual features with RS captions as language typically contains more precise semantic than visual data. Furthermore, to address the challenges posed by high similarity in RS datasets, we introduce an effective hybrid cross-modal loss that facilitates comprehensive multimodal feature alignment within a unified structure. We conduct extensive experiments on three benchmark datasets, achieving state-of-the-art performance, which validates the effectiveness and superiority of our method.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"18 ","pages":"10172-10185"},"PeriodicalIF":4.7000,"publicationDate":"2025-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10945411","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10945411/","RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Remote sensing (RS) image–text matching has gained significant attention for its promising potential. Despite great advancements, accurately matching RS images (RSIs) and captions remains challenging due to the significant multimodal gap and inherent characteristics of RS data. Many approaches use complex models to extract global features to handle semantic redundancy and varying scales in RSIs, but losing important details in RSIs and captions. While some methods align between fine-grained local features, but overlooking the semantic granularity differences between fine-grained features. Fine-grained features in RSIs typically capture only a small fraction of the overall semantics, whereas those in captions convey more comprehensive and abstract semantics. Therefore, we propose the prompt-based granularity-unified representation network, an end-to-end framework designed to mitigate the multimodal semantic granularity difference and achieve comprehensive alignment. Our approach includes two key modules: 1) the prompt-based feature aggregator, which dynamically aggregates fine-grained features into several granularity-unified tokens with fully semantic, and 2) the text-guided vision modulation, which further enhances visual representations by modulating the visual features with RS captions as language typically contains more precise semantic than visual data. Furthermore, to address the challenges posed by high similarity in RS datasets, we introduce an effective hybrid cross-modal loss that facilitates comprehensive multimodal feature alignment within a unified structure. We conduct extensive experiments on three benchmark datasets, achieving state-of-the-art performance, which validates the effectiveness and superiority of our method.
期刊介绍:
The IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing addresses the growing field of applications in Earth observations and remote sensing, and also provides a venue for the rapidly expanding special issues that are being sponsored by the IEEE Geosciences and Remote Sensing Society. The journal draws upon the experience of the highly successful “IEEE Transactions on Geoscience and Remote Sensing” and provide a complementary medium for the wide range of topics in applied earth observations. The ‘Applications’ areas encompasses the societal benefit areas of the Global Earth Observations Systems of Systems (GEOSS) program. Through deliberations over two years, ministers from 50 countries agreed to identify nine areas where Earth observation could positively impact the quality of life and health of their respective countries. Some of these are areas not traditionally addressed in the IEEE context. These include biodiversity, health and climate. Yet it is the skill sets of IEEE members, in areas such as observations, communications, computers, signal processing, standards and ocean engineering, that form the technical underpinnings of GEOSS. Thus, the Journal attracts a broad range of interests that serves both present members in new ways and expands the IEEE visibility into new areas.