Zvezdan Loncarevic, Mihael Simonič, A. Ude, A. Gams
{"title":"Combining Reinforcement Learning and Lazy Learning for Faster Few-Shot Transfer Learning","authors":"Zvezdan Loncarevic, Mihael Simonič, A. Ude, A. Gams","doi":"10.1109/Humanoids53995.2022.10000095","DOIUrl":null,"url":null,"abstract":"Since repeating a task with a humanoid robot many times is typically far too time consuming and strenuous for the robotic mechanism, learning is often shifted to simulation. Bridging the sim-to-real gap, however, still requires considerable real-world effort. In this paper we explore how to reduce the number of required repetitions with a novel few-shot transfer learning methodology. The skill is initially encoded with a deep neural network in one domain, and later adapted for a different target domain by re-training only a partllayer of this deep neural network with real data. For retraining we propose to combine lazy learning and reinforcement learning. Our experiments show that such combination is considerably faster than only using either one of these and an order of magnitude faster than learning from scratch. We demonstrated the approach on the example of robotic throwing, a complex dynamic skill where the outcome of the task is not explicitly dependent on the final position of the robot motion. The experiments were performed for sim-to-sim transfer learning on the full-sized humanoid robot TALOS, with initial throwing implementation on the real robot.","PeriodicalId":180816,"journal":{"name":"2022 IEEE-RAS 21st International Conference on Humanoid Robots (Humanoids)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE-RAS 21st International Conference on Humanoid Robots (Humanoids)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/Humanoids53995.2022.10000095","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Since repeating a task with a humanoid robot many times is typically far too time consuming and strenuous for the robotic mechanism, learning is often shifted to simulation. Bridging the sim-to-real gap, however, still requires considerable real-world effort. In this paper we explore how to reduce the number of required repetitions with a novel few-shot transfer learning methodology. The skill is initially encoded with a deep neural network in one domain, and later adapted for a different target domain by re-training only a partllayer of this deep neural network with real data. For retraining we propose to combine lazy learning and reinforcement learning. Our experiments show that such combination is considerably faster than only using either one of these and an order of magnitude faster than learning from scratch. We demonstrated the approach on the example of robotic throwing, a complex dynamic skill where the outcome of the task is not explicitly dependent on the final position of the robot motion. The experiments were performed for sim-to-sim transfer learning on the full-sized humanoid robot TALOS, with initial throwing implementation on the real robot.