{"title":"基于h -∞滤波和强化学习算法的城市环境智能导航","authors":"Ivan Smolyakov, R. Langley","doi":"10.1109/PLANS46316.2020.9109948","DOIUrl":null,"url":null,"abstract":"In urban areas, robustness of a positioning solution suffers from relatively unpredictable reception of attenuated, non-line-of-sight and multipath-contaminated signals. To reflect a GNSS signal propagation environment, parameters of a state estimation filter need to be adjusted on-the-fly. A mixed H2/ H∞ filter has been considered here to address the vulnerability of a minimum error variance estimator to measurement outliers. An emphasis between the H2filter and the H∞ filter (minimizing the worst-case error) is continuously adjusted by a reinforcement learning (RL) model. Specifically, a continuous action actor-critic RL model with eligibility traces is implemented. The Cramér-Rao lower bound is considered for the filter performance evaluation allowing for the RL reward computation. The algorithm has been tested on a real-world dataset collected with mass-market hardware applying tightly-coupled IMU/GPS sensor integration. A positive RL model learning trend has been identified in two segments of the trajectory with the highest obstruction environment, suggesting the applicability potential of the technique.","PeriodicalId":273568,"journal":{"name":"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Intelligent Navigation in Urban Environments Based on an H-infinity Filter and Reinforcement Learning Algorithms\",\"authors\":\"Ivan Smolyakov, R. Langley\",\"doi\":\"10.1109/PLANS46316.2020.9109948\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In urban areas, robustness of a positioning solution suffers from relatively unpredictable reception of attenuated, non-line-of-sight and multipath-contaminated signals. To reflect a GNSS signal propagation environment, parameters of a state estimation filter need to be adjusted on-the-fly. A mixed H2/ H∞ filter has been considered here to address the vulnerability of a minimum error variance estimator to measurement outliers. An emphasis between the H2filter and the H∞ filter (minimizing the worst-case error) is continuously adjusted by a reinforcement learning (RL) model. Specifically, a continuous action actor-critic RL model with eligibility traces is implemented. The Cramér-Rao lower bound is considered for the filter performance evaluation allowing for the RL reward computation. The algorithm has been tested on a real-world dataset collected with mass-market hardware applying tightly-coupled IMU/GPS sensor integration. A positive RL model learning trend has been identified in two segments of the trajectory with the highest obstruction environment, suggesting the applicability potential of the technique.\",\"PeriodicalId\":273568,\"journal\":{\"name\":\"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)\",\"volume\":\"43 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/PLANS46316.2020.9109948\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE/ION Position, Location and Navigation Symposium (PLANS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PLANS46316.2020.9109948","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Intelligent Navigation in Urban Environments Based on an H-infinity Filter and Reinforcement Learning Algorithms
In urban areas, robustness of a positioning solution suffers from relatively unpredictable reception of attenuated, non-line-of-sight and multipath-contaminated signals. To reflect a GNSS signal propagation environment, parameters of a state estimation filter need to be adjusted on-the-fly. A mixed H2/ H∞ filter has been considered here to address the vulnerability of a minimum error variance estimator to measurement outliers. An emphasis between the H2filter and the H∞ filter (minimizing the worst-case error) is continuously adjusted by a reinforcement learning (RL) model. Specifically, a continuous action actor-critic RL model with eligibility traces is implemented. The Cramér-Rao lower bound is considered for the filter performance evaluation allowing for the RL reward computation. The algorithm has been tested on a real-world dataset collected with mass-market hardware applying tightly-coupled IMU/GPS sensor integration. A positive RL model learning trend has been identified in two segments of the trajectory with the highest obstruction environment, suggesting the applicability potential of the technique.