Mazbahur Rahman Khan, Azhar Mohd Ibrahim, Suaib Al Mahmud, Farah Asyiqin Samat, Farahiyah Jasni, Muhammad Imran Mardzuki
{"title":"基于DRL和启发式奖励的移动机器人导航技术综述","authors":"Mazbahur Rahman Khan, Azhar Mohd Ibrahim, Suaib Al Mahmud, Farah Asyiqin Samat, Farahiyah Jasni, Muhammad Imran Mardzuki","doi":"10.1016/j.neucom.2025.131036","DOIUrl":null,"url":null,"abstract":"<div><div>Robotic navigation is a critical component of autonomy, requiring efficient and safe mobility across diverse environments. The advent of Deep Reinforcement Learning (DRL) has spurred significant research into enabling mobile robots to learn effective navigation by optimizing actions based on environmental rewards. DRL has shown promise in addressing challenges such as dynamic environments and cooperative exploration. However, traditional DRL-based navigation faces several limitations, including the need for extensive training data, susceptibility to local traps in complex environments, low transferability to real-world scenarios, slow convergence, and low learning efficiency. Additionally, designing an appropriate reward function to achieve desired behaviors without unintended consequences remains complex; poorly designed rewards can lead to suboptimal or harmful outcomes. Recent studies have explored integrating heuristic search-based rewards into DRL algorithms to mitigate these issues. This study reviews the limitations of traditional DRL navigation and explores recent advancements in integrating heuristic search to design dynamic reward functions that enhance robot learning processes.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"652 ","pages":"Article 131036"},"PeriodicalIF":5.5000,"publicationDate":"2025-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Advancing mobile robot navigation with DRL and heuristic rewards: A comprehensive review\",\"authors\":\"Mazbahur Rahman Khan, Azhar Mohd Ibrahim, Suaib Al Mahmud, Farah Asyiqin Samat, Farahiyah Jasni, Muhammad Imran Mardzuki\",\"doi\":\"10.1016/j.neucom.2025.131036\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Robotic navigation is a critical component of autonomy, requiring efficient and safe mobility across diverse environments. The advent of Deep Reinforcement Learning (DRL) has spurred significant research into enabling mobile robots to learn effective navigation by optimizing actions based on environmental rewards. DRL has shown promise in addressing challenges such as dynamic environments and cooperative exploration. However, traditional DRL-based navigation faces several limitations, including the need for extensive training data, susceptibility to local traps in complex environments, low transferability to real-world scenarios, slow convergence, and low learning efficiency. Additionally, designing an appropriate reward function to achieve desired behaviors without unintended consequences remains complex; poorly designed rewards can lead to suboptimal or harmful outcomes. Recent studies have explored integrating heuristic search-based rewards into DRL algorithms to mitigate these issues. This study reviews the limitations of traditional DRL navigation and explores recent advancements in integrating heuristic search to design dynamic reward functions that enhance robot learning processes.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"652 \",\"pages\":\"Article 131036\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-07-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225017084\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225017084","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Advancing mobile robot navigation with DRL and heuristic rewards: A comprehensive review
Robotic navigation is a critical component of autonomy, requiring efficient and safe mobility across diverse environments. The advent of Deep Reinforcement Learning (DRL) has spurred significant research into enabling mobile robots to learn effective navigation by optimizing actions based on environmental rewards. DRL has shown promise in addressing challenges such as dynamic environments and cooperative exploration. However, traditional DRL-based navigation faces several limitations, including the need for extensive training data, susceptibility to local traps in complex environments, low transferability to real-world scenarios, slow convergence, and low learning efficiency. Additionally, designing an appropriate reward function to achieve desired behaviors without unintended consequences remains complex; poorly designed rewards can lead to suboptimal or harmful outcomes. Recent studies have explored integrating heuristic search-based rewards into DRL algorithms to mitigate these issues. This study reviews the limitations of traditional DRL navigation and explores recent advancements in integrating heuristic search to design dynamic reward functions that enhance robot learning processes.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.