Autonomous Navigation and Collision Avoidance for AGV in Dynamic Environments: An Enhanced Deep Reinforcement Learning Approach With Composite Rewards and Dynamic Update Mechanisms
{"title":"Autonomous Navigation and Collision Avoidance for AGV in Dynamic Environments: An Enhanced Deep Reinforcement Learning Approach With Composite Rewards and Dynamic Update Mechanisms","authors":"Zijianglong Huang, Zhigang Ren, Tehuan Chen, Shengze Cai, Chao Xu","doi":"10.1049/csy2.70012","DOIUrl":null,"url":null,"abstract":"<p>With the booming development of logistics, manufacturing and warehousing fields, the autonomous navigation and intelligent obstacle avoidance technology of automated guided vehicles (AGVs) has become the focus of scientific research. In this paper, an enhanced deep reinforcement learning (DRL) framework is proposed, aiming to empower AGVs with the ability of autonomous navigation and obstacle avoidance in the unknown and variable complex environment. To address the problems of time-consuming training and limited generalisation ability of traditional DRL, we refine the twin delayed deep deterministic policy gradient algorithm by integrating adaptive noise attenuation and dynamic delayed updating, optimising both training efficiency and model robustness. In order to further strengthen the AGV's ability to perceive and respond to changes of a dynamic environment, we introduce a distance-based obstacle penalty term in the designed composite reward function, which ensures that the AGV is capable of predicting and avoiding obstacles effectively in dynamic scenarios. Experiments indicate that the AGV model trained by this algorithm presents excellent autonomous navigation capability in both static and dynamic environments, with a high task completion rate, stable and reliable operation, which fully proves the high efficiency and robustness of this method and its practical value.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70012","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Cybersystems and Robotics","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/csy2.70012","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
With the booming development of logistics, manufacturing and warehousing fields, the autonomous navigation and intelligent obstacle avoidance technology of automated guided vehicles (AGVs) has become the focus of scientific research. In this paper, an enhanced deep reinforcement learning (DRL) framework is proposed, aiming to empower AGVs with the ability of autonomous navigation and obstacle avoidance in the unknown and variable complex environment. To address the problems of time-consuming training and limited generalisation ability of traditional DRL, we refine the twin delayed deep deterministic policy gradient algorithm by integrating adaptive noise attenuation and dynamic delayed updating, optimising both training efficiency and model robustness. In order to further strengthen the AGV's ability to perceive and respond to changes of a dynamic environment, we introduce a distance-based obstacle penalty term in the designed composite reward function, which ensures that the AGV is capable of predicting and avoiding obstacles effectively in dynamic scenarios. Experiments indicate that the AGV model trained by this algorithm presents excellent autonomous navigation capability in both static and dynamic environments, with a high task completion rate, stable and reliable operation, which fully proves the high efficiency and robustness of this method and its practical value.