Zhen Wang , Dianxi Shi , Chunping Qiu , Songchang Jin , Tongyue Li , Ziteng Qiao , Yang Chen
{"title":"VecMapLocNet: Vision-based UAV localization using vector maps in GNSS-denied environments","authors":"Zhen Wang , Dianxi Shi , Chunping Qiu , Songchang Jin , Tongyue Li , Ziteng Qiao , Yang Chen","doi":"10.1016/j.isprsjprs.2025.04.009","DOIUrl":null,"url":null,"abstract":"<div><div>Vision-based localization techniques are effective UAV localization solutions for GNSS-denied conditions, however they depend on costly, complex, and seasonally variable satellite images or 3D maps, whereas humans can determine location using vector maps. Inspired by human navigation strategies, we propose VecMapLocNet , which uses vector maps to determine UAV 3-DoF poses (latitude, longitude, and yaw) through cross-modal matching. Three key modules are designed to improve the matching between UAV images and vector maps. The UAV feature extraction module is low-latency and adaptable to various flight altitudes, ensuring it is suitable for airborne deployment. The vector map feature extraction module employs a weighted representation of different map elements, ensuring robustness against changes in visual appearance. Inspired by Fourier transforms, the feature matching module for 3-DoF pose estimation is parameter-free, computationally efficient, and invariant to cross-modal differences. To evaluate VecMapLocNet, we introduce a comprehensive dataset that presents challenges, encompassing seven cities worldwide. Through rigorous experimentation, VecMapLocNet has demonstrated competitive performance compared to existing methods in localization accuracy (84.45% Recall@5 m), yaw estimation (88.61% Recall@5°), and computational efficiency (25.23ms latency on onboard device Jetson Orin). Furthermore, we validated VecMapLocNet’s performance in real-world scenarios, with experimental results confirming its generalization ability, achieving a localization error of 16.7 m and an orientation error of 3.1°. The code and datasets are available at <span><span>https://map.geovisuallocalization.com</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"225 ","pages":"Pages 362-381"},"PeriodicalIF":10.6000,"publicationDate":"2025-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625001455","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Vision-based localization techniques are effective UAV localization solutions for GNSS-denied conditions, however they depend on costly, complex, and seasonally variable satellite images or 3D maps, whereas humans can determine location using vector maps. Inspired by human navigation strategies, we propose VecMapLocNet , which uses vector maps to determine UAV 3-DoF poses (latitude, longitude, and yaw) through cross-modal matching. Three key modules are designed to improve the matching between UAV images and vector maps. The UAV feature extraction module is low-latency and adaptable to various flight altitudes, ensuring it is suitable for airborne deployment. The vector map feature extraction module employs a weighted representation of different map elements, ensuring robustness against changes in visual appearance. Inspired by Fourier transforms, the feature matching module for 3-DoF pose estimation is parameter-free, computationally efficient, and invariant to cross-modal differences. To evaluate VecMapLocNet, we introduce a comprehensive dataset that presents challenges, encompassing seven cities worldwide. Through rigorous experimentation, VecMapLocNet has demonstrated competitive performance compared to existing methods in localization accuracy (84.45% Recall@5 m), yaw estimation (88.61% Recall@5°), and computational efficiency (25.23ms latency on onboard device Jetson Orin). Furthermore, we validated VecMapLocNet’s performance in real-world scenarios, with experimental results confirming its generalization ability, achieving a localization error of 16.7 m and an orientation error of 3.1°. The code and datasets are available at https://map.geovisuallocalization.com.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.