Mapless navigation via Hierarchical Reinforcement Learning with memory-decaying novelty

IF 4.3 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS
Yan Gao , Feiqiang Lin , Boliang Cai , Jing Wu , Changyun Wei , Raphael Grech , Ze Ji
{"title":"Mapless navigation via Hierarchical Reinforcement Learning with memory-decaying novelty","authors":"Yan Gao ,&nbsp;Feiqiang Lin ,&nbsp;Boliang Cai ,&nbsp;Jing Wu ,&nbsp;Changyun Wei ,&nbsp;Raphael Grech ,&nbsp;Ze Ji","doi":"10.1016/j.robot.2024.104815","DOIUrl":null,"url":null,"abstract":"<div><div>Hierarchical Reinforcement Learning (HRL) has shown superior performance for mapless navigation tasks. However, it remains limited in unstructured environments that might contain terrains like long corridors and dead corners, which can lead to local minima. This is because most HRL-based mapless navigation methods employ a simplified reward setting and exploration strategy. In this work, we propose a novel reward function for training the high-level (HL) policy, which contains two components: extrinsic reward and intrinsic reward. The extrinsic reward encourages the robot to move towards the target location, while the intrinsic reward is computed based on novelty, episode memory and memory decaying, making the agent capable of accomplishing spontaneous exploration. We also design a novel neural network structure that incorporates an LSTM network to augment the agent with memory and reasoning capabilities. We test our method in unknown environments and specific scenarios prone to the local minimum problem to evaluate the navigation performance and local minimum resolution ability. The results show that our method significantly increases the success rate when compared to advanced RL-based methods, achieving a maximum improvement of nearly 28%. Our method demonstrates effective improvement in addressing the local minimum issue, especially in cases where the baselines fail completely. Additionally, numerous ablation studies consistently confirm the effectiveness of our proposed reward function and neural network structure.</div></div>","PeriodicalId":49592,"journal":{"name":"Robotics and Autonomous Systems","volume":"182 ","pages":"Article 104815"},"PeriodicalIF":4.3000,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0921889024001994/pdfft?md5=d8582d1a21a5b1405a794b5c34b147fa&pid=1-s2.0-S0921889024001994-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Robotics and Autonomous Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0921889024001994","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Hierarchical Reinforcement Learning (HRL) has shown superior performance for mapless navigation tasks. However, it remains limited in unstructured environments that might contain terrains like long corridors and dead corners, which can lead to local minima. This is because most HRL-based mapless navigation methods employ a simplified reward setting and exploration strategy. In this work, we propose a novel reward function for training the high-level (HL) policy, which contains two components: extrinsic reward and intrinsic reward. The extrinsic reward encourages the robot to move towards the target location, while the intrinsic reward is computed based on novelty, episode memory and memory decaying, making the agent capable of accomplishing spontaneous exploration. We also design a novel neural network structure that incorporates an LSTM network to augment the agent with memory and reasoning capabilities. We test our method in unknown environments and specific scenarios prone to the local minimum problem to evaluate the navigation performance and local minimum resolution ability. The results show that our method significantly increases the success rate when compared to advanced RL-based methods, achieving a maximum improvement of nearly 28%. Our method demonstrates effective improvement in addressing the local minimum issue, especially in cases where the baselines fail completely. Additionally, numerous ablation studies consistently confirm the effectiveness of our proposed reward function and neural network structure.
通过具有记忆衰减新奇感的分层强化学习实现无地图导航
分层强化学习(HRL)在无地图导航任务中表现出卓越的性能。然而,在可能包含长走廊和死角等地形的非结构化环境中,它仍然受到限制,因为这些地形可能导致局部最小值。这是因为大多数基于 HRL 的无地图导航方法都采用了简化的奖励设置和探索策略。在这项工作中,我们提出了一种用于训练高级(HL)策略的新型奖励函数,它包含两个部分:外在奖励和内在奖励。外在奖励鼓励机器人向目标位置移动,而内在奖励则根据新颖性、情节记忆和记忆衰减来计算,从而使机器人能够完成自发探索。我们还设计了一种新颖的神经网络结构,其中包含一个 LSTM 网络,以增强机器人的记忆和推理能力。我们在未知环境和容易出现局部最小值问题的特定场景中测试了我们的方法,以评估导航性能和局部最小值的解决能力。结果表明,与先进的基于 RL 的方法相比,我们的方法大大提高了成功率,最大提高了近 28%。我们的方法有效改善了局部最小值问题的解决,尤其是在基线完全失效的情况下。此外,大量的消融研究一致证实了我们提出的奖励函数和神经网络结构的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Robotics and Autonomous Systems
Robotics and Autonomous Systems 工程技术-机器人学
CiteScore
9.00
自引率
7.00%
发文量
164
审稿时长
4.5 months
期刊介绍: Robotics and Autonomous Systems will carry articles describing fundamental developments in the field of robotics, with special emphasis on autonomous systems. An important goal of this journal is to extend the state of the art in both symbolic and sensory based robot control and learning in the context of autonomous systems. Robotics and Autonomous Systems will carry articles on the theoretical, computational and experimental aspects of autonomous systems, or modules of such systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信