{"title":"基于层次深度强化学习的机器人避障运动规划","authors":"Guoquan Zhao, Fen Ying, Zuowei Pang, Huashan Liu","doi":"10.1145/3598151.3598170","DOIUrl":null,"url":null,"abstract":"When the task environment becomes complex, deep reinforcement learning (DRL) is easy to encounter the problems of gradient disappearance or explosion. To solve this problem, this paper proposes a hierarchical DRL framework consisting of task and action layers. The task layer learns interpretable representations of tasks and decision processes, and drives the action layer. The action layer learns to collaboratively accomplish complex tasks in different roles. The DRL algorithm based on this framework is tested on a redundant degree of freedom robot in obstacle avoidance motion planning tasks, and comparative experimental results prove the effectiveness and feasibility of the proposed method.","PeriodicalId":398644,"journal":{"name":"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Robotic motion planning with obstacle avoidance based on hierarchical deep reinforcement learning\",\"authors\":\"Guoquan Zhao, Fen Ying, Zuowei Pang, Huashan Liu\",\"doi\":\"10.1145/3598151.3598170\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"When the task environment becomes complex, deep reinforcement learning (DRL) is easy to encounter the problems of gradient disappearance or explosion. To solve this problem, this paper proposes a hierarchical DRL framework consisting of task and action layers. The task layer learns interpretable representations of tasks and decision processes, and drives the action layer. The action layer learns to collaboratively accomplish complex tasks in different roles. The DRL algorithm based on this framework is tested on a redundant degree of freedom robot in obstacle avoidance motion planning tasks, and comparative experimental results prove the effectiveness and feasibility of the proposed method.\",\"PeriodicalId\":398644,\"journal\":{\"name\":\"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering\",\"volume\":\"14 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3598151.3598170\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3598151.3598170","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Robotic motion planning with obstacle avoidance based on hierarchical deep reinforcement learning
When the task environment becomes complex, deep reinforcement learning (DRL) is easy to encounter the problems of gradient disappearance or explosion. To solve this problem, this paper proposes a hierarchical DRL framework consisting of task and action layers. The task layer learns interpretable representations of tasks and decision processes, and drives the action layer. The action layer learns to collaboratively accomplish complex tasks in different roles. The DRL algorithm based on this framework is tested on a redundant degree of freedom robot in obstacle avoidance motion planning tasks, and comparative experimental results prove the effectiveness and feasibility of the proposed method.