{"title":"移动机器人导航的深度强化学习","authors":"M. Gromniak, Jonas Stenzel","doi":"10.1109/ACIRS.2019.8935944","DOIUrl":null,"url":null,"abstract":"While navigation is arguable the most important aspect of mobile robotics, complex scenarios with dynamic environments or with teams of cooperative robots are still not satisfactory solved yet. Motivated by the recent successes in the reinforcement learning domain, the application of deep reinforcement learning to robot navigation was examined in this paper. In particular this required the development of a training procedure, a set of actions available to the robot, a suitable state representation and a reward function. The setup was evaluated using a simulated real-time environment. A reference setup, different goal-oriented exploration strategies and two different robot kinematics (holonomic, differential) were compared in the evaluation. In a challenging scenario with obstacles at changing locations in the environment the robot was able to reach the desired goal in 93% of the episodes.","PeriodicalId":338050,"journal":{"name":"2019 4th Asia-Pacific Conference on Intelligent Robot Systems (ACIRS)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Deep Reinforcement Learning for Mobile Robot Navigation\",\"authors\":\"M. Gromniak, Jonas Stenzel\",\"doi\":\"10.1109/ACIRS.2019.8935944\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"While navigation is arguable the most important aspect of mobile robotics, complex scenarios with dynamic environments or with teams of cooperative robots are still not satisfactory solved yet. Motivated by the recent successes in the reinforcement learning domain, the application of deep reinforcement learning to robot navigation was examined in this paper. In particular this required the development of a training procedure, a set of actions available to the robot, a suitable state representation and a reward function. The setup was evaluated using a simulated real-time environment. A reference setup, different goal-oriented exploration strategies and two different robot kinematics (holonomic, differential) were compared in the evaluation. In a challenging scenario with obstacles at changing locations in the environment the robot was able to reach the desired goal in 93% of the episodes.\",\"PeriodicalId\":338050,\"journal\":{\"name\":\"2019 4th Asia-Pacific Conference on Intelligent Robot Systems (ACIRS)\",\"volume\":\"58 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 4th Asia-Pacific Conference on Intelligent Robot Systems (ACIRS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ACIRS.2019.8935944\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 4th Asia-Pacific Conference on Intelligent Robot Systems (ACIRS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACIRS.2019.8935944","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep Reinforcement Learning for Mobile Robot Navigation
While navigation is arguable the most important aspect of mobile robotics, complex scenarios with dynamic environments or with teams of cooperative robots are still not satisfactory solved yet. Motivated by the recent successes in the reinforcement learning domain, the application of deep reinforcement learning to robot navigation was examined in this paper. In particular this required the development of a training procedure, a set of actions available to the robot, a suitable state representation and a reward function. The setup was evaluated using a simulated real-time environment. A reference setup, different goal-oriented exploration strategies and two different robot kinematics (holonomic, differential) were compared in the evaluation. In a challenging scenario with obstacles at changing locations in the environment the robot was able to reach the desired goal in 93% of the episodes.