Xianyong Ruan, Du Jiang, Juntong Yun, Bo Tao, Yuanmin Xie, Baojia Chen, Meng Jia, Li Huang
{"title":"Improved DDPG-Based Path Planning for Mobile Robots","authors":"Xianyong Ruan, Du Jiang, Juntong Yun, Bo Tao, Yuanmin Xie, Baojia Chen, Meng Jia, Li Huang","doi":"10.1002/cpe.70317","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>With the rapid advancement of robotics technology, path planning has attracted extensive research attention. Reinforcement learning, owing to its ability to acquire optimal policies through continuous interaction with the environment, offers a promising solution for path planning in environments with incomplete or unknown information. However, reinforcement learning-based path planning methods often suffer from high training complexity and low utilization of effective samples. To address these issues, this paper proposes an improved deep reinforcement learning (DRL) algorithm. The proposed approach builds upon the deep deterministic policy gradient (DDPG) algorithm and incorporates a short-term goal planning strategy based on local perceptual information, which decomposes the global navigation task into multiple short-term subgoals, thereby reducing task complexity and enhancing learning efficiency. Furthermore, a reward function integrating the artificial potential field (APF) method is designed to improve obstacle avoidance capability. To tackle the low utilization of effective experiences in DDPG, a dual experience pool strategy is introduced to improve experience utilization efficiency and accelerate model training. The parameters for short-term goal selection are optimized through multiple comparative experiments, and the proposed method is evaluated against several DRL-based path planning approaches in a static environment. Experimental results demonstrate that the improved algorithm significantly accelerates convergence. Moreover, dynamic environment simulation experiments verify that the proposed algorithm can effectively avoid moving obstacles and achieve safe navigation to the target position.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"37 25-26","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.70317","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
With the rapid advancement of robotics technology, path planning has attracted extensive research attention. Reinforcement learning, owing to its ability to acquire optimal policies through continuous interaction with the environment, offers a promising solution for path planning in environments with incomplete or unknown information. However, reinforcement learning-based path planning methods often suffer from high training complexity and low utilization of effective samples. To address these issues, this paper proposes an improved deep reinforcement learning (DRL) algorithm. The proposed approach builds upon the deep deterministic policy gradient (DDPG) algorithm and incorporates a short-term goal planning strategy based on local perceptual information, which decomposes the global navigation task into multiple short-term subgoals, thereby reducing task complexity and enhancing learning efficiency. Furthermore, a reward function integrating the artificial potential field (APF) method is designed to improve obstacle avoidance capability. To tackle the low utilization of effective experiences in DDPG, a dual experience pool strategy is introduced to improve experience utilization efficiency and accelerate model training. The parameters for short-term goal selection are optimized through multiple comparative experiments, and the proposed method is evaluated against several DRL-based path planning approaches in a static environment. Experimental results demonstrate that the improved algorithm significantly accelerates convergence. Moreover, dynamic environment simulation experiments verify that the proposed algorithm can effectively avoid moving obstacles and achieve safe navigation to the target position.
期刊介绍:
Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of:
Parallel and distributed computing;
High-performance computing;
Computational and data science;
Artificial intelligence and machine learning;
Big data applications, algorithms, and systems;
Network science;
Ontologies and semantics;
Security and privacy;
Cloud/edge/fog computing;
Green computing; and
Quantum computing.