基于深度强化学习的不确定动态环境下自主移动机器人导航

Zhangfan Lu, Ran Huang
{"title":"基于深度强化学习的不确定动态环境下自主移动机器人导航","authors":"Zhangfan Lu, Ran Huang","doi":"10.1109/RCAR52367.2021.9517635","DOIUrl":null,"url":null,"abstract":"In this paper, we study autonomous end-to-end navigation for wheeled robots based on deep reinforcement learning (DRL) in an unknown environment without a priori map. The DRL network is mainly based on deep deterministic policy gradient algorithm together with long short-term memory. The input for the network is the data from a 2D lidar as well as the relative position to the target point, while the outputs are the linear velocity and angular velocity that actuate the robot. A novel reward function is proposed to avoid the collision with dynamic obstacles and to generate a smooth trajectory for the robot. The network is trained without supervision in an unknown dynamic environment, the random Gaussian noise is added to the input data of long short-term memory to avoid local optimum. Besides, different unstructured environments are also considered in the training to increase the robustness of the developed network. Experiments performed on public dataset have showed that the developed network makes the robot navigate in unstructured environments safely and outperform several DRL methods.","PeriodicalId":232892,"journal":{"name":"2021 IEEE International Conference on Real-time Computing and Robotics (RCAR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Autonomous mobile robot navigation in uncertain dynamic environments based on deep reinforcement learning\",\"authors\":\"Zhangfan Lu, Ran Huang\",\"doi\":\"10.1109/RCAR52367.2021.9517635\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we study autonomous end-to-end navigation for wheeled robots based on deep reinforcement learning (DRL) in an unknown environment without a priori map. The DRL network is mainly based on deep deterministic policy gradient algorithm together with long short-term memory. The input for the network is the data from a 2D lidar as well as the relative position to the target point, while the outputs are the linear velocity and angular velocity that actuate the robot. A novel reward function is proposed to avoid the collision with dynamic obstacles and to generate a smooth trajectory for the robot. The network is trained without supervision in an unknown dynamic environment, the random Gaussian noise is added to the input data of long short-term memory to avoid local optimum. Besides, different unstructured environments are also considered in the training to increase the robustness of the developed network. Experiments performed on public dataset have showed that the developed network makes the robot navigate in unstructured environments safely and outperform several DRL methods.\",\"PeriodicalId\":232892,\"journal\":{\"name\":\"2021 IEEE International Conference on Real-time Computing and Robotics (RCAR)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Real-time Computing and Robotics (RCAR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/RCAR52367.2021.9517635\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Real-time Computing and Robotics (RCAR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/RCAR52367.2021.9517635","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

在本文中,我们研究了基于深度强化学习(DRL)的轮式机器人在未知环境中没有先验地图的自主端到端导航。DRL网络主要基于深度确定性策略梯度算法和长短期记忆。网络的输入是来自2D激光雷达的数据以及与目标点的相对位置,而输出是驱动机器人的线速度和角速度。提出了一种新的奖励函数,以避免与动态障碍物的碰撞,并使机器人产生平滑的运动轨迹。该网络在未知的动态环境下进行无监督训练,在长短期记忆的输入数据中加入随机高斯噪声以避免局部最优。此外,在训练中还考虑了不同的非结构化环境,以增加所开发网络的鲁棒性。在公共数据集上进行的实验表明,所开发的网络使机器人在非结构化环境中安全导航,并且优于几种DRL方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Autonomous mobile robot navigation in uncertain dynamic environments based on deep reinforcement learning
In this paper, we study autonomous end-to-end navigation for wheeled robots based on deep reinforcement learning (DRL) in an unknown environment without a priori map. The DRL network is mainly based on deep deterministic policy gradient algorithm together with long short-term memory. The input for the network is the data from a 2D lidar as well as the relative position to the target point, while the outputs are the linear velocity and angular velocity that actuate the robot. A novel reward function is proposed to avoid the collision with dynamic obstacles and to generate a smooth trajectory for the robot. The network is trained without supervision in an unknown dynamic environment, the random Gaussian noise is added to the input data of long short-term memory to avoid local optimum. Besides, different unstructured environments are also considered in the training to increase the robustness of the developed network. Experiments performed on public dataset have showed that the developed network makes the robot navigate in unstructured environments safely and outperform several DRL methods.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信