基于改进强化学习优化的移动机器人路径规划

Yanshu Jing, Yukun Chen, Ming-hai Jiao, Jie Huang, Bowen Niu, Wenbo Zheng
{"title":"基于改进强化学习优化的移动机器人路径规划","authors":"Yanshu Jing, Yukun Chen, Ming-hai Jiao, Jie Huang, Bowen Niu, Wenbo Zheng","doi":"10.1145/3366715.3366717","DOIUrl":null,"url":null,"abstract":"The constant parameter is usually set in adaptive function with traditional mobile robot path planning problem. Q-learning, a type of reinforcement learning, has gained increasing popularity in autonomous mobile robot path recently. In order to effectively solve mobile robot path planning problem in obstacle avoidance environment, a path planning model and search algorithm based on improved reinforcement learning are proposed. The incentive model of reinforcement learning mechanism is introduced with search selection strategy, modifying dynamic reward function parameter setting. The group intelligent search iterative process of global position selection and local position selection is exploited to combine particle behavior with reinforcement learning algorithm, dynamically adjusting the empirical parameter of the reward function by strengthening the data training experiment of Q-learning. to determine the constant parameters for simulation experiment, once the distance between the robot and the obstacle is less than a certain thresholds value, the 0-1 random number is used to randomly adjust the moving direction, avoiding the occurrence of mobile robot path matching deadlock. The study case shows that the proposed algorithm is proved to be better efficient and effective, thereby improving the search intensity and accuracy of the mobile robot path planning problem. And the experimental simulation shows that the proposed model and algorithm effectively solve mobile robot path planning problem that the parameter selection and the actual scene cannot be adapted in real time in traditional path planning problem.","PeriodicalId":425980,"journal":{"name":"Proceedings of the 2019 International Conference on Robotics Systems and Vehicle Technology - RSVT '19","volume":"53 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Mobile Robot Path Planning Based on Improved Reinforcement Learning Optimization\",\"authors\":\"Yanshu Jing, Yukun Chen, Ming-hai Jiao, Jie Huang, Bowen Niu, Wenbo Zheng\",\"doi\":\"10.1145/3366715.3366717\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The constant parameter is usually set in adaptive function with traditional mobile robot path planning problem. Q-learning, a type of reinforcement learning, has gained increasing popularity in autonomous mobile robot path recently. In order to effectively solve mobile robot path planning problem in obstacle avoidance environment, a path planning model and search algorithm based on improved reinforcement learning are proposed. The incentive model of reinforcement learning mechanism is introduced with search selection strategy, modifying dynamic reward function parameter setting. The group intelligent search iterative process of global position selection and local position selection is exploited to combine particle behavior with reinforcement learning algorithm, dynamically adjusting the empirical parameter of the reward function by strengthening the data training experiment of Q-learning. to determine the constant parameters for simulation experiment, once the distance between the robot and the obstacle is less than a certain thresholds value, the 0-1 random number is used to randomly adjust the moving direction, avoiding the occurrence of mobile robot path matching deadlock. The study case shows that the proposed algorithm is proved to be better efficient and effective, thereby improving the search intensity and accuracy of the mobile robot path planning problem. And the experimental simulation shows that the proposed model and algorithm effectively solve mobile robot path planning problem that the parameter selection and the actual scene cannot be adapted in real time in traditional path planning problem.\",\"PeriodicalId\":425980,\"journal\":{\"name\":\"Proceedings of the 2019 International Conference on Robotics Systems and Vehicle Technology - RSVT '19\",\"volume\":\"53 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2019 International Conference on Robotics Systems and Vehicle Technology - RSVT '19\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3366715.3366717\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2019 International Conference on Robotics Systems and Vehicle Technology - RSVT '19","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3366715.3366717","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

传统的移动机器人路径规划问题通常在自适应函数中设置常数参数。q学习作为强化学习的一种,近年来在自主移动机器人路径中得到了越来越广泛的应用。为了有效解决避障环境下移动机器人的路径规划问题,提出了一种基于改进强化学习的路径规划模型和搜索算法。引入了基于搜索选择策略的强化学习机制激励模型,修改了动态奖励函数的参数设置。利用全局位置选择和局部位置选择的群体智能搜索迭代过程,将粒子行为与强化学习算法相结合,通过强化Q-learning的数据训练实验,动态调整奖励函数的经验参数。为确定仿真实验的恒定参数,一旦机器人与障碍物的距离小于某一阈值,采用0-1随机数随机调整移动方向,避免移动机器人路径匹配死锁的发生。研究实例表明,该算法具有较高的效率和有效性,从而提高了移动机器人路径规划问题的搜索强度和精度。实验仿真表明,所提出的模型和算法有效地解决了传统路径规划问题中参数选择和实际场景不能实时适应的移动机器人路径规划问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Mobile Robot Path Planning Based on Improved Reinforcement Learning Optimization
The constant parameter is usually set in adaptive function with traditional mobile robot path planning problem. Q-learning, a type of reinforcement learning, has gained increasing popularity in autonomous mobile robot path recently. In order to effectively solve mobile robot path planning problem in obstacle avoidance environment, a path planning model and search algorithm based on improved reinforcement learning are proposed. The incentive model of reinforcement learning mechanism is introduced with search selection strategy, modifying dynamic reward function parameter setting. The group intelligent search iterative process of global position selection and local position selection is exploited to combine particle behavior with reinforcement learning algorithm, dynamically adjusting the empirical parameter of the reward function by strengthening the data training experiment of Q-learning. to determine the constant parameters for simulation experiment, once the distance between the robot and the obstacle is less than a certain thresholds value, the 0-1 random number is used to randomly adjust the moving direction, avoiding the occurrence of mobile robot path matching deadlock. The study case shows that the proposed algorithm is proved to be better efficient and effective, thereby improving the search intensity and accuracy of the mobile robot path planning problem. And the experimental simulation shows that the proposed model and algorithm effectively solve mobile robot path planning problem that the parameter selection and the actual scene cannot be adapted in real time in traditional path planning problem.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信