{"title":"基于样本高效强化学习的移动机器人姿态调节","authors":"Walter Brescia, L. D. Cicco, S. Mascolo","doi":"10.1109/ICCAIS56082.2022.9990480","DOIUrl":null,"url":null,"abstract":"Reinforcement Learning (RL) has gained interest in the control and automation communities thanks to its encouraging results in many challenging control problems without requiring a model of the system and of the environment. Yet, it is well-known that employing such a learning-based approach in real scenarios may be problematic, as a prohibitive amount of data might be required to converge to an optimal control policy. In this work, we equip a popular RL algorithm with two tools to improve exploration effectiveness and sample efficiency: the Episodic Noise, that helps useful subsets of actions emerge already in the first few training episodes, and the Difficulty Manager, that generates goals proportioned to the current agent’s capabilities. We demonstrate the effectiveness of such proposed tools on a pose regulation task of a four wheel steering four wheel driving robot, suitable for a wide range of industrial scenarios. The resulting agent learns effective sets of actions in just a few hundreds training epochs, reaching satisfactory performance during tests.","PeriodicalId":273404,"journal":{"name":"2022 11th International Conference on Control, Automation and Information Sciences (ICCAIS)","volume":"212 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Sample-Efficient Reinforcement Learning for Pose Regulation of a Mobile Robot\",\"authors\":\"Walter Brescia, L. D. Cicco, S. Mascolo\",\"doi\":\"10.1109/ICCAIS56082.2022.9990480\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Reinforcement Learning (RL) has gained interest in the control and automation communities thanks to its encouraging results in many challenging control problems without requiring a model of the system and of the environment. Yet, it is well-known that employing such a learning-based approach in real scenarios may be problematic, as a prohibitive amount of data might be required to converge to an optimal control policy. In this work, we equip a popular RL algorithm with two tools to improve exploration effectiveness and sample efficiency: the Episodic Noise, that helps useful subsets of actions emerge already in the first few training episodes, and the Difficulty Manager, that generates goals proportioned to the current agent’s capabilities. We demonstrate the effectiveness of such proposed tools on a pose regulation task of a four wheel steering four wheel driving robot, suitable for a wide range of industrial scenarios. The resulting agent learns effective sets of actions in just a few hundreds training epochs, reaching satisfactory performance during tests.\",\"PeriodicalId\":273404,\"journal\":{\"name\":\"2022 11th International Conference on Control, Automation and Information Sciences (ICCAIS)\",\"volume\":\"212 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 11th International Conference on Control, Automation and Information Sciences (ICCAIS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCAIS56082.2022.9990480\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 11th International Conference on Control, Automation and Information Sciences (ICCAIS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCAIS56082.2022.9990480","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Sample-Efficient Reinforcement Learning for Pose Regulation of a Mobile Robot
Reinforcement Learning (RL) has gained interest in the control and automation communities thanks to its encouraging results in many challenging control problems without requiring a model of the system and of the environment. Yet, it is well-known that employing such a learning-based approach in real scenarios may be problematic, as a prohibitive amount of data might be required to converge to an optimal control policy. In this work, we equip a popular RL algorithm with two tools to improve exploration effectiveness and sample efficiency: the Episodic Noise, that helps useful subsets of actions emerge already in the first few training episodes, and the Difficulty Manager, that generates goals proportioned to the current agent’s capabilities. We demonstrate the effectiveness of such proposed tools on a pose regulation task of a four wheel steering four wheel driving robot, suitable for a wide range of industrial scenarios. The resulting agent learns effective sets of actions in just a few hundreds training epochs, reaching satisfactory performance during tests.