{"title":"Cross-Modal Localization: Using automotive radar for absolute geolocation within a map produced with visible-light imagery","authors":"Peter A. Iannucci, Lakshay Narula, T. Humphreys","doi":"10.1109/PLANS46316.2020.9110143","DOIUrl":null,"url":null,"abstract":"This paper explores the possibility of localizing an automotive-radar-equipped vehicle within an urban environment relative to an existing map of the environment created using data from visible light cameras. Such cross-modal localization would enable robust, low-cost absolute localization in poor weather conditions based only on radar even when the vehicle has never previously visited the area. This is because a pre-existing absolutely-referenced visible-light-based map (e.g., constructed from Google Street View images) could be exploited for localization provided that a correspondence between features in this map and the vehicle's radar returns can be established. The greatest challenge presented by cross-modal localization with automotive radar is the extreme sparseness of automotive-radar-produced features, which prevents application of standard computer vision techniques for the cross-modal registration. To the best of the authors' knowledge, cross-modal localization using automotive-grade radar within a visible-light-based map is unprecedented. The current paper demonstrates that it can be used for vehicle localization with horizontal errors below 61 cm (95%).","PeriodicalId":273568,"journal":{"name":"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PLANS46316.2020.9110143","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
This paper explores the possibility of localizing an automotive-radar-equipped vehicle within an urban environment relative to an existing map of the environment created using data from visible light cameras. Such cross-modal localization would enable robust, low-cost absolute localization in poor weather conditions based only on radar even when the vehicle has never previously visited the area. This is because a pre-existing absolutely-referenced visible-light-based map (e.g., constructed from Google Street View images) could be exploited for localization provided that a correspondence between features in this map and the vehicle's radar returns can be established. The greatest challenge presented by cross-modal localization with automotive radar is the extreme sparseness of automotive-radar-produced features, which prevents application of standard computer vision techniques for the cross-modal registration. To the best of the authors' knowledge, cross-modal localization using automotive-grade radar within a visible-light-based map is unprecedented. The current paper demonstrates that it can be used for vehicle localization with horizontal errors below 61 cm (95%).