{"title":"基于连续动作空间的深度强化学习拓扑信息编码","authors":"Qi Wang, Yue Gao, Wei Liu","doi":"10.1109/ICRAE48301.2019.9043778","DOIUrl":null,"url":null,"abstract":"In the context of reinforcement learning, the training efficiency can decay exponentially with the size of the state space. Therefore, designing easily-optimized state space representation has remained an open problem. In this paper, we focus on a general and challenging scenario, i.e. reinforcement learning with continuous action spaces. We propose a new representation framework by explicitly encoding topology information such as the geometrical and the kinematic relations among different parts of the agent to make the representation more informative, which results in effective optimization. Extensive experiments were conducted on three settings to demonstrate that our method can remarkably stabilize and speed up the training process.","PeriodicalId":270665,"journal":{"name":"2019 4th International Conference on Robotics and Automation Engineering (ICRAE)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Encoding Topology Information for Deep Reinforcement Learning with Continuous Action Space\",\"authors\":\"Qi Wang, Yue Gao, Wei Liu\",\"doi\":\"10.1109/ICRAE48301.2019.9043778\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the context of reinforcement learning, the training efficiency can decay exponentially with the size of the state space. Therefore, designing easily-optimized state space representation has remained an open problem. In this paper, we focus on a general and challenging scenario, i.e. reinforcement learning with continuous action spaces. We propose a new representation framework by explicitly encoding topology information such as the geometrical and the kinematic relations among different parts of the agent to make the representation more informative, which results in effective optimization. Extensive experiments were conducted on three settings to demonstrate that our method can remarkably stabilize and speed up the training process.\",\"PeriodicalId\":270665,\"journal\":{\"name\":\"2019 4th International Conference on Robotics and Automation Engineering (ICRAE)\",\"volume\":\"11 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 4th International Conference on Robotics and Automation Engineering (ICRAE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICRAE48301.2019.9043778\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 4th International Conference on Robotics and Automation Engineering (ICRAE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICRAE48301.2019.9043778","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Encoding Topology Information for Deep Reinforcement Learning with Continuous Action Space
In the context of reinforcement learning, the training efficiency can decay exponentially with the size of the state space. Therefore, designing easily-optimized state space representation has remained an open problem. In this paper, we focus on a general and challenging scenario, i.e. reinforcement learning with continuous action spaces. We propose a new representation framework by explicitly encoding topology information such as the geometrical and the kinematic relations among different parts of the agent to make the representation more informative, which results in effective optimization. Extensive experiments were conducted on three settings to demonstrate that our method can remarkably stabilize and speed up the training process.