基于DRL和启发式奖励的移动机器人导航技术综述

IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Mazbahur Rahman Khan, Azhar Mohd Ibrahim, Suaib Al Mahmud, Farah Asyiqin Samat, Farahiyah Jasni, Muhammad Imran Mardzuki
{"title":"基于DRL和启发式奖励的移动机器人导航技术综述","authors":"Mazbahur Rahman Khan,&nbsp;Azhar Mohd Ibrahim,&nbsp;Suaib Al Mahmud,&nbsp;Farah Asyiqin Samat,&nbsp;Farahiyah Jasni,&nbsp;Muhammad Imran Mardzuki","doi":"10.1016/j.neucom.2025.131036","DOIUrl":null,"url":null,"abstract":"<div><div>Robotic navigation is a critical component of autonomy, requiring efficient and safe mobility across diverse environments. The advent of Deep Reinforcement Learning (DRL) has spurred significant research into enabling mobile robots to learn effective navigation by optimizing actions based on environmental rewards. DRL has shown promise in addressing challenges such as dynamic environments and cooperative exploration. However, traditional DRL-based navigation faces several limitations, including the need for extensive training data, susceptibility to local traps in complex environments, low transferability to real-world scenarios, slow convergence, and low learning efficiency. Additionally, designing an appropriate reward function to achieve desired behaviors without unintended consequences remains complex; poorly designed rewards can lead to suboptimal or harmful outcomes. Recent studies have explored integrating heuristic search-based rewards into DRL algorithms to mitigate these issues. This study reviews the limitations of traditional DRL navigation and explores recent advancements in integrating heuristic search to design dynamic reward functions that enhance robot learning processes.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"652 ","pages":"Article 131036"},"PeriodicalIF":5.5000,"publicationDate":"2025-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Advancing mobile robot navigation with DRL and heuristic rewards: A comprehensive review\",\"authors\":\"Mazbahur Rahman Khan,&nbsp;Azhar Mohd Ibrahim,&nbsp;Suaib Al Mahmud,&nbsp;Farah Asyiqin Samat,&nbsp;Farahiyah Jasni,&nbsp;Muhammad Imran Mardzuki\",\"doi\":\"10.1016/j.neucom.2025.131036\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Robotic navigation is a critical component of autonomy, requiring efficient and safe mobility across diverse environments. The advent of Deep Reinforcement Learning (DRL) has spurred significant research into enabling mobile robots to learn effective navigation by optimizing actions based on environmental rewards. DRL has shown promise in addressing challenges such as dynamic environments and cooperative exploration. However, traditional DRL-based navigation faces several limitations, including the need for extensive training data, susceptibility to local traps in complex environments, low transferability to real-world scenarios, slow convergence, and low learning efficiency. Additionally, designing an appropriate reward function to achieve desired behaviors without unintended consequences remains complex; poorly designed rewards can lead to suboptimal or harmful outcomes. Recent studies have explored integrating heuristic search-based rewards into DRL algorithms to mitigate these issues. This study reviews the limitations of traditional DRL navigation and explores recent advancements in integrating heuristic search to design dynamic reward functions that enhance robot learning processes.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"652 \",\"pages\":\"Article 131036\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-07-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225017084\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225017084","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

机器人导航是自动驾驶的关键组成部分,需要在各种环境中高效、安全的移动。深度强化学习(DRL)的出现刺激了大量研究,使移动机器人能够通过基于环境奖励优化行动来学习有效的导航。DRL在应对诸如动态环境和合作探索等挑战方面显示出了希望。然而,传统的基于drl的导航面临着一些限制,包括需要大量的训练数据,在复杂环境中容易受到局部陷阱的影响,对现实场景的可移植性低,收敛速度慢,学习效率低。此外,设计适当的奖励功能以实现预期行为而不产生意外后果仍然很复杂;设计不佳的奖励会导致次优或有害的结果。最近的研究探索了将基于启发式搜索的奖励集成到DRL算法中,以缓解这些问题。本研究回顾了传统DRL导航的局限性,并探讨了集成启发式搜索来设计动态奖励函数以增强机器人学习过程的最新进展。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Advancing mobile robot navigation with DRL and heuristic rewards: A comprehensive review
Robotic navigation is a critical component of autonomy, requiring efficient and safe mobility across diverse environments. The advent of Deep Reinforcement Learning (DRL) has spurred significant research into enabling mobile robots to learn effective navigation by optimizing actions based on environmental rewards. DRL has shown promise in addressing challenges such as dynamic environments and cooperative exploration. However, traditional DRL-based navigation faces several limitations, including the need for extensive training data, susceptibility to local traps in complex environments, low transferability to real-world scenarios, slow convergence, and low learning efficiency. Additionally, designing an appropriate reward function to achieve desired behaviors without unintended consequences remains complex; poorly designed rewards can lead to suboptimal or harmful outcomes. Recent studies have explored integrating heuristic search-based rewards into DRL algorithms to mitigate these issues. This study reviews the limitations of traditional DRL navigation and explores recent advancements in integrating heuristic search to design dynamic reward functions that enhance robot learning processes.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信