Xiong Qiu, Shouyi Liao, Dongfang Yang, Yongfei Li, Shicheng Wang
{"title":"Visual geo-localization and attitude estimation using satellite imagery and topographical elevation for unmanned aerial vehicles","authors":"Xiong Qiu, Shouyi Liao, Dongfang Yang, Yongfei Li, Shicheng Wang","doi":"10.1016/j.engappai.2025.110759","DOIUrl":null,"url":null,"abstract":"<div><div>In recent years, the autonomous flight capability of unmanned aerial vehicles (UAVs) has significantly improved. However, basic issues such as geo-localization and orientation still remain unsolved when Global Positioning System (GPS) is unavailable. Current visual geo-localization methods either assume a flat flight area using only satellite imagery for large-scale scenes or rely on three-dimensional (3D) reconstruction for small-scale areas, both of which limit practical UAVs applications. To achieve visual geo-localization in large-scale scenes with complex terrain variations, we propose a method that reliably achieves 6-degree-of-freedom pose estimation, including offline database and online inference. In the first stage, neural networks extract global and local features from satellite images to build feature databases. In the second stage, a fast search is conducted in the global feature database based on the global image descriptor of the aerial image, followed by fine matching based on local keypoint descriptors. Geographic coordinates for matched feature points are provided by satellite imagery and topographical elevation to achieve the UAV pose estimation. We conducted experimental validation by capturing aerial images at two non-overlapping locations. The results show that at Location 1, when using a top-down view, the overall localization error of the UAV decreased from 31.76 m to 17.94 m, improving accuracy by 43.61%; the recall rate increased from 67.59% to 71.94%. At Location 2, when using a front-down view, the overall localization error of the UAV decreased from 48.07 m to 10.84 m, with accuracy significantly improving by 77.45%; the recall rate increased from 64.96% to 69.71%.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"153 ","pages":"Article 110759"},"PeriodicalIF":8.0000,"publicationDate":"2025-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625007596","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
In recent years, the autonomous flight capability of unmanned aerial vehicles (UAVs) has significantly improved. However, basic issues such as geo-localization and orientation still remain unsolved when Global Positioning System (GPS) is unavailable. Current visual geo-localization methods either assume a flat flight area using only satellite imagery for large-scale scenes or rely on three-dimensional (3D) reconstruction for small-scale areas, both of which limit practical UAVs applications. To achieve visual geo-localization in large-scale scenes with complex terrain variations, we propose a method that reliably achieves 6-degree-of-freedom pose estimation, including offline database and online inference. In the first stage, neural networks extract global and local features from satellite images to build feature databases. In the second stage, a fast search is conducted in the global feature database based on the global image descriptor of the aerial image, followed by fine matching based on local keypoint descriptors. Geographic coordinates for matched feature points are provided by satellite imagery and topographical elevation to achieve the UAV pose estimation. We conducted experimental validation by capturing aerial images at two non-overlapping locations. The results show that at Location 1, when using a top-down view, the overall localization error of the UAV decreased from 31.76 m to 17.94 m, improving accuracy by 43.61%; the recall rate increased from 67.59% to 71.94%. At Location 2, when using a front-down view, the overall localization error of the UAV decreased from 48.07 m to 10.84 m, with accuracy significantly improving by 77.45%; the recall rate increased from 64.96% to 69.71%.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.