{"title":"Sample-Efficient Reinforcement Learning for Pose Regulation of a Mobile Robot","authors":"Walter Brescia, L. D. Cicco, S. Mascolo","doi":"10.1109/ICCAIS56082.2022.9990480","DOIUrl":null,"url":null,"abstract":"Reinforcement Learning (RL) has gained interest in the control and automation communities thanks to its encouraging results in many challenging control problems without requiring a model of the system and of the environment. Yet, it is well-known that employing such a learning-based approach in real scenarios may be problematic, as a prohibitive amount of data might be required to converge to an optimal control policy. In this work, we equip a popular RL algorithm with two tools to improve exploration effectiveness and sample efficiency: the Episodic Noise, that helps useful subsets of actions emerge already in the first few training episodes, and the Difficulty Manager, that generates goals proportioned to the current agent’s capabilities. We demonstrate the effectiveness of such proposed tools on a pose regulation task of a four wheel steering four wheel driving robot, suitable for a wide range of industrial scenarios. The resulting agent learns effective sets of actions in just a few hundreds training epochs, reaching satisfactory performance during tests.","PeriodicalId":273404,"journal":{"name":"2022 11th International Conference on Control, Automation and Information Sciences (ICCAIS)","volume":"212 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 11th International Conference on Control, Automation and Information Sciences (ICCAIS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCAIS56082.2022.9990480","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Reinforcement Learning (RL) has gained interest in the control and automation communities thanks to its encouraging results in many challenging control problems without requiring a model of the system and of the environment. Yet, it is well-known that employing such a learning-based approach in real scenarios may be problematic, as a prohibitive amount of data might be required to converge to an optimal control policy. In this work, we equip a popular RL algorithm with two tools to improve exploration effectiveness and sample efficiency: the Episodic Noise, that helps useful subsets of actions emerge already in the first few training episodes, and the Difficulty Manager, that generates goals proportioned to the current agent’s capabilities. We demonstrate the effectiveness of such proposed tools on a pose regulation task of a four wheel steering four wheel driving robot, suitable for a wide range of industrial scenarios. The resulting agent learns effective sets of actions in just a few hundreds training epochs, reaching satisfactory performance during tests.