{"title":"Intelligent Navigation in Urban Environments Based on an H-infinity Filter and Reinforcement Learning Algorithms","authors":"Ivan Smolyakov, R. Langley","doi":"10.1109/PLANS46316.2020.9109948","DOIUrl":null,"url":null,"abstract":"In urban areas, robustness of a positioning solution suffers from relatively unpredictable reception of attenuated, non-line-of-sight and multipath-contaminated signals. To reflect a GNSS signal propagation environment, parameters of a state estimation filter need to be adjusted on-the-fly. A mixed H2/ H∞ filter has been considered here to address the vulnerability of a minimum error variance estimator to measurement outliers. An emphasis between the H2filter and the H∞ filter (minimizing the worst-case error) is continuously adjusted by a reinforcement learning (RL) model. Specifically, a continuous action actor-critic RL model with eligibility traces is implemented. The Cramér-Rao lower bound is considered for the filter performance evaluation allowing for the RL reward computation. The algorithm has been tested on a real-world dataset collected with mass-market hardware applying tightly-coupled IMU/GPS sensor integration. A positive RL model learning trend has been identified in two segments of the trajectory with the highest obstruction environment, suggesting the applicability potential of the technique.","PeriodicalId":273568,"journal":{"name":"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PLANS46316.2020.9109948","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
In urban areas, robustness of a positioning solution suffers from relatively unpredictable reception of attenuated, non-line-of-sight and multipath-contaminated signals. To reflect a GNSS signal propagation environment, parameters of a state estimation filter need to be adjusted on-the-fly. A mixed H2/ H∞ filter has been considered here to address the vulnerability of a minimum error variance estimator to measurement outliers. An emphasis between the H2filter and the H∞ filter (minimizing the worst-case error) is continuously adjusted by a reinforcement learning (RL) model. Specifically, a continuous action actor-critic RL model with eligibility traces is implemented. The Cramér-Rao lower bound is considered for the filter performance evaluation allowing for the RL reward computation. The algorithm has been tested on a real-world dataset collected with mass-market hardware applying tightly-coupled IMU/GPS sensor integration. A positive RL model learning trend has been identified in two segments of the trajectory with the highest obstruction environment, suggesting the applicability potential of the technique.