{"title":"基于深度强化学习的密集行人环境下机器人三维仿真环境与导航方法","authors":"Qi Liu, Yanjie Li, Lintao Liu","doi":"10.1109/CASE48305.2020.9217023","DOIUrl":null,"url":null,"abstract":"With the rapid development of mobile robot technology, robots are playing an increasingly important role in people’s daily lives. As one of the key technologies of the basic functions of mobile robots, navigation also needs to deal with new challenges. How to navigate efficiently and collision-free in complex and changeable human environments is one of the problems that need to be solved urgently. Currently, mobile robots can achieve efficient navigation in static environments. However, in the unstructured and fast-changing environments of human daily society, robots need to make more flexible navigation strategies to deal with the dynamic scenarios. This paper built a 3D simulation environment for robot navigation via deep reinforcement learning in dense pedestrian environment. We also proposed a new navigation approach via deep reinforcement learning in dense pedestrian environment. The simulation environment of this paper integrates Gazebo, ROS navigation stack, Stable baselines and Social Force Pedestrian Simulator. In order to be able to collect the rich environmental information around the robot, our simulation environment is based on the Gazebo simulation platform. In order to use the traditional path planning methods, we introduce the ROS navigation stack. In order to make it easier to call the current mainstream reinforcement learning algorithms, we introduce Stable baselines which is a set of improved implementations of reinforcement learning algorithms based on OpenAI Baselines. In order to imitate dense pedestrian scenarios realistically, we introduce the Social Force Pedestrian Simulator which is a pedestrian simulation package, whose pedestrian’s movement follows the rules of Social Force Movement. Our robot navigation approach combines the global optimality of traditional global path planning and the local barrier ability of reinforcement learning. Firstly, we plan global path by using A* algorithm. Secondly, we use Soft Actor Critic (SAC) to try to follow the waypoints generated at a certain distance on the global path to make action decisions on the premise of agile obstacle avoidance. Experiments show that our simulation environment can easily set up a robot navigation environment and navigation approaches can be simulated in various dense pedestrian environments.","PeriodicalId":212181,"journal":{"name":"2020 IEEE 16th International Conference on Automation Science and Engineering (CASE)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"A 3D Simulation Environment and Navigation Approach for Robot Navigation via Deep Reinforcement Learning in Dense Pedestrian Environment\",\"authors\":\"Qi Liu, Yanjie Li, Lintao Liu\",\"doi\":\"10.1109/CASE48305.2020.9217023\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the rapid development of mobile robot technology, robots are playing an increasingly important role in people’s daily lives. As one of the key technologies of the basic functions of mobile robots, navigation also needs to deal with new challenges. How to navigate efficiently and collision-free in complex and changeable human environments is one of the problems that need to be solved urgently. Currently, mobile robots can achieve efficient navigation in static environments. However, in the unstructured and fast-changing environments of human daily society, robots need to make more flexible navigation strategies to deal with the dynamic scenarios. This paper built a 3D simulation environment for robot navigation via deep reinforcement learning in dense pedestrian environment. We also proposed a new navigation approach via deep reinforcement learning in dense pedestrian environment. The simulation environment of this paper integrates Gazebo, ROS navigation stack, Stable baselines and Social Force Pedestrian Simulator. In order to be able to collect the rich environmental information around the robot, our simulation environment is based on the Gazebo simulation platform. In order to use the traditional path planning methods, we introduce the ROS navigation stack. In order to make it easier to call the current mainstream reinforcement learning algorithms, we introduce Stable baselines which is a set of improved implementations of reinforcement learning algorithms based on OpenAI Baselines. In order to imitate dense pedestrian scenarios realistically, we introduce the Social Force Pedestrian Simulator which is a pedestrian simulation package, whose pedestrian’s movement follows the rules of Social Force Movement. Our robot navigation approach combines the global optimality of traditional global path planning and the local barrier ability of reinforcement learning. Firstly, we plan global path by using A* algorithm. Secondly, we use Soft Actor Critic (SAC) to try to follow the waypoints generated at a certain distance on the global path to make action decisions on the premise of agile obstacle avoidance. Experiments show that our simulation environment can easily set up a robot navigation environment and navigation approaches can be simulated in various dense pedestrian environments.\",\"PeriodicalId\":212181,\"journal\":{\"name\":\"2020 IEEE 16th International Conference on Automation Science and Engineering (CASE)\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE 16th International Conference on Automation Science and Engineering (CASE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CASE48305.2020.9217023\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE 16th International Conference on Automation Science and Engineering (CASE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CASE48305.2020.9217023","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
摘要
随着移动机器人技术的飞速发展,机器人在人们的日常生活中扮演着越来越重要的角色。导航作为移动机器人基本功能的关键技术之一,也需要应对新的挑战。如何在复杂多变的人类环境中高效、无碰撞地进行导航是目前迫切需要解决的问题之一。目前,移动机器人可以在静态环境中实现高效导航。然而,在人类日常社会的非结构化和快速变化的环境中,机器人需要制定更灵活的导航策略来应对动态场景。本文通过深度强化学习,构建了密集行人环境下机器人导航的三维仿真环境。我们还提出了一种新的基于深度强化学习的密集行人环境下的导航方法。本文的仿真环境集成了Gazebo、ROS导航堆栈、Stable基线和Social Force Pedestrian Simulator。为了能够采集机器人周围丰富的环境信息,我们的仿真环境基于Gazebo仿真平台。为了沿用传统的路径规划方法,我们引入了ROS导航堆栈。为了方便调用当前主流的强化学习算法,我们引入了Stable baselines,这是一组基于OpenAI baselines的强化学习算法的改进实现。为了逼真地模拟密集的行人场景,我们引入了Social Force pedestrian Simulator,这是一个行人仿真包,行人的运动遵循Social Force的运动规则。我们的机器人导航方法结合了传统全局路径规划的全局最优性和强化学习的局部障碍能力。首先,采用A*算法规划全局路径;其次,在敏捷避障的前提下,利用软行为批评家(Soft Actor Critic, SAC)尝试在全局路径上一定距离处生成的路径点进行行动决策。实验表明,我们的仿真环境可以很容易地建立一个机器人导航环境,并且可以在各种密集的行人环境中模拟导航方法。
A 3D Simulation Environment and Navigation Approach for Robot Navigation via Deep Reinforcement Learning in Dense Pedestrian Environment
With the rapid development of mobile robot technology, robots are playing an increasingly important role in people’s daily lives. As one of the key technologies of the basic functions of mobile robots, navigation also needs to deal with new challenges. How to navigate efficiently and collision-free in complex and changeable human environments is one of the problems that need to be solved urgently. Currently, mobile robots can achieve efficient navigation in static environments. However, in the unstructured and fast-changing environments of human daily society, robots need to make more flexible navigation strategies to deal with the dynamic scenarios. This paper built a 3D simulation environment for robot navigation via deep reinforcement learning in dense pedestrian environment. We also proposed a new navigation approach via deep reinforcement learning in dense pedestrian environment. The simulation environment of this paper integrates Gazebo, ROS navigation stack, Stable baselines and Social Force Pedestrian Simulator. In order to be able to collect the rich environmental information around the robot, our simulation environment is based on the Gazebo simulation platform. In order to use the traditional path planning methods, we introduce the ROS navigation stack. In order to make it easier to call the current mainstream reinforcement learning algorithms, we introduce Stable baselines which is a set of improved implementations of reinforcement learning algorithms based on OpenAI Baselines. In order to imitate dense pedestrian scenarios realistically, we introduce the Social Force Pedestrian Simulator which is a pedestrian simulation package, whose pedestrian’s movement follows the rules of Social Force Movement. Our robot navigation approach combines the global optimality of traditional global path planning and the local barrier ability of reinforcement learning. Firstly, we plan global path by using A* algorithm. Secondly, we use Soft Actor Critic (SAC) to try to follow the waypoints generated at a certain distance on the global path to make action decisions on the premise of agile obstacle avoidance. Experiments show that our simulation environment can easily set up a robot navigation environment and navigation approaches can be simulated in various dense pedestrian environments.