Deep Reinforcement Learning for Localisability-Aware Mapless Navigation

IF 1.5 Q3 AUTOMATION & CONTROL SYSTEMS
Yan Gao, Jing Wu, Changyun Wei, Raphael Grech, Ze Ji
{"title":"Deep Reinforcement Learning for Localisability-Aware Mapless Navigation","authors":"Yan Gao,&nbsp;Jing Wu,&nbsp;Changyun Wei,&nbsp;Raphael Grech,&nbsp;Ze Ji","doi":"10.1049/csy2.70018","DOIUrl":null,"url":null,"abstract":"<p>Mapless navigation refers to the task of searching for a collision free path without relying on a pre-defined map. Most current works of mapless navigation assume accurate ground-truth localisation is available. However, this is not true, especially for indoor environments, where simultaneous localisation and mapping (SLAM) is needed for location estimation, which highly relies on the richness of environment features. In this work, we propose a novel deep reinforcement learning (DRL) based mapless navigation method without relying on the assumption of the availability of localisation. Our method utilises RGB-D based Oriented FAST and Rotated BRIEF (ORB) SLAM2 for robot localisation. Our policy effectively guides the robot's movement towards the target while enhancing robot pose estimation by considering the quality of the observed features along the selected paths. To facilitate policy training, we propose a compact state representation based on the spatial distributions of map points, which enhances the robot's awareness of areas with reliable map points. Furthermore, we suggest incorporating the relative pose error into the reward function. In this way, the policy will be more responsive to each single action. In addition, rather than utilising a pre-set threshold, we adopt a dynamic threshold to improve the policy's adaptability to variations in SLAM performance across different environments. The experiments in localisation challenging environments have demonstrated the remarkable performance of our proposed method. It outperforms the related DRL based methods in terms of success rate.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70018","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Cybersystems and Robotics","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/csy2.70018","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Mapless navigation refers to the task of searching for a collision free path without relying on a pre-defined map. Most current works of mapless navigation assume accurate ground-truth localisation is available. However, this is not true, especially for indoor environments, where simultaneous localisation and mapping (SLAM) is needed for location estimation, which highly relies on the richness of environment features. In this work, we propose a novel deep reinforcement learning (DRL) based mapless navigation method without relying on the assumption of the availability of localisation. Our method utilises RGB-D based Oriented FAST and Rotated BRIEF (ORB) SLAM2 for robot localisation. Our policy effectively guides the robot's movement towards the target while enhancing robot pose estimation by considering the quality of the observed features along the selected paths. To facilitate policy training, we propose a compact state representation based on the spatial distributions of map points, which enhances the robot's awareness of areas with reliable map points. Furthermore, we suggest incorporating the relative pose error into the reward function. In this way, the policy will be more responsive to each single action. In addition, rather than utilising a pre-set threshold, we adopt a dynamic threshold to improve the policy's adaptability to variations in SLAM performance across different environments. The experiments in localisation challenging environments have demonstrated the remarkable performance of our proposed method. It outperforms the related DRL based methods in terms of success rate.

Abstract Image

基于可定位性感知的无地图导航深度强化学习
无地图导航是指在不依赖于预定义地图的情况下搜索无碰撞路径的任务。目前大多数无地图导航工作都假定可以获得精确的地真定位。然而,情况并非如此,特别是在室内环境中,位置估计需要同时定位和映射(SLAM),这高度依赖于环境特征的丰富性。在这项工作中,我们提出了一种新的基于深度强化学习(DRL)的无地图导航方法,而不依赖于定位可用性的假设。我们的方法利用基于RGB-D的定向FAST和旋转BRIEF (ORB) SLAM2进行机器人定位。我们的策略有效地引导机器人向目标移动,同时通过考虑在所选路径上观察到的特征的质量来增强机器人的姿态估计。为了便于策略训练,我们提出了一种基于地图点空间分布的紧凑状态表示,增强了机器人对具有可靠地图点区域的感知。此外,我们建议将相对姿态误差纳入奖励函数。通过这种方式,策略对每个单个操作的响应将更加灵敏。此外,我们采用动态阈值来提高策略对不同环境中SLAM性能变化的适应性,而不是使用预先设置的阈值。在具有定位挑战性的环境中进行的实验证明了我们提出的方法的显著性能。在成功率方面优于相关的基于DRL的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IET Cybersystems and Robotics
IET Cybersystems and Robotics Computer Science-Information Systems
CiteScore
3.70
自引率
0.00%
发文量
31
审稿时长
34 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信