Xubo Luo, Yaolin Tian, Xue Wan, Jingzhong Xu, Tao Ke
{"title":"基于深度学习的无人机地理定位横视图像匹配","authors":"Xubo Luo, Yaolin Tian, Xue Wan, Jingzhong Xu, Tao Ke","doi":"10.1109/ICoSR57188.2022.00028","DOIUrl":null,"url":null,"abstract":"Unmanned Aerial Vehicles (UAVs) geo-localization refers to finding the position of a given aerial image in a large reference satellite image. Due to the large scale and illumination difference between aerial and satellite images, it is challenging that most existing cross-view image matching algorithms fail to localize the UAV images robustly and accurately. To solve the above problem, a novel UAV localization framework containing three-stage coarse-to-fine image matching is proposed. In the first stage, the satellite image is cropped into several local reference images to be matched with the aerial image. Then, ten candidate local images are selected from all of the local reference images with a simple and effective deep learning network, LPN. At last, a deep feature-based matching is employed between candidate local reference images and aerial images to determine the optimal position of the UAV in the reference map via homography transformation. In addition, a satellite-UAV image dataset is proposed, which contains 3 large-scale satellite images and 1909 aerial images. To demonstrate the performance of the proposed method, experiments on the large-scale proposed dataset are conducted. The experimental results illustrate that for more than 80% of the testing pair images, the proposed method is capable of refining the positioning error within 5 pixels, which meets the needs of UAV localization and is superior to other popular methods.","PeriodicalId":234590,"journal":{"name":"2022 International Conference on Service Robotics (ICoSR)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep learning based cross-view image matching for UAV geo-localization\",\"authors\":\"Xubo Luo, Yaolin Tian, Xue Wan, Jingzhong Xu, Tao Ke\",\"doi\":\"10.1109/ICoSR57188.2022.00028\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Unmanned Aerial Vehicles (UAVs) geo-localization refers to finding the position of a given aerial image in a large reference satellite image. Due to the large scale and illumination difference between aerial and satellite images, it is challenging that most existing cross-view image matching algorithms fail to localize the UAV images robustly and accurately. To solve the above problem, a novel UAV localization framework containing three-stage coarse-to-fine image matching is proposed. In the first stage, the satellite image is cropped into several local reference images to be matched with the aerial image. Then, ten candidate local images are selected from all of the local reference images with a simple and effective deep learning network, LPN. At last, a deep feature-based matching is employed between candidate local reference images and aerial images to determine the optimal position of the UAV in the reference map via homography transformation. In addition, a satellite-UAV image dataset is proposed, which contains 3 large-scale satellite images and 1909 aerial images. To demonstrate the performance of the proposed method, experiments on the large-scale proposed dataset are conducted. The experimental results illustrate that for more than 80% of the testing pair images, the proposed method is capable of refining the positioning error within 5 pixels, which meets the needs of UAV localization and is superior to other popular methods.\",\"PeriodicalId\":234590,\"journal\":{\"name\":\"2022 International Conference on Service Robotics (ICoSR)\",\"volume\":\"55 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 International Conference on Service Robotics (ICoSR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICoSR57188.2022.00028\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Service Robotics (ICoSR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICoSR57188.2022.00028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep learning based cross-view image matching for UAV geo-localization
Unmanned Aerial Vehicles (UAVs) geo-localization refers to finding the position of a given aerial image in a large reference satellite image. Due to the large scale and illumination difference between aerial and satellite images, it is challenging that most existing cross-view image matching algorithms fail to localize the UAV images robustly and accurately. To solve the above problem, a novel UAV localization framework containing three-stage coarse-to-fine image matching is proposed. In the first stage, the satellite image is cropped into several local reference images to be matched with the aerial image. Then, ten candidate local images are selected from all of the local reference images with a simple and effective deep learning network, LPN. At last, a deep feature-based matching is employed between candidate local reference images and aerial images to determine the optimal position of the UAV in the reference map via homography transformation. In addition, a satellite-UAV image dataset is proposed, which contains 3 large-scale satellite images and 1909 aerial images. To demonstrate the performance of the proposed method, experiments on the large-scale proposed dataset are conducted. The experimental results illustrate that for more than 80% of the testing pair images, the proposed method is capable of refining the positioning error within 5 pixels, which meets the needs of UAV localization and is superior to other popular methods.