Nícolas F. Figueroa;Julio C. Tafur;Abderrahmane Kheddar
{"title":"Fast Autolearning for Multimodal Walking in Humanoid Robots With Variability of Experience","authors":"Nícolas F. Figueroa;Julio C. Tafur;Abderrahmane Kheddar","doi":"10.1109/LRA.2025.3546168","DOIUrl":null,"url":null,"abstract":"Recent advancements in reinforcement learning (RL) and humanoid robotics are rapidly addressing the challenge of adapting to complex, dynamic environments in real time. This letter introduces a novel approach that integrates two key concepts: experience variability (a criterion for detecting changes in loco-manipulation) and experience accumulation (an efficient method for storing acquired experiences based on a selection criterion). These elements are incorporated into the development of RL agents and humanoid robots, with an emphasis on stability. This focus enhances adaptability and efficiency in unpredictable environments. Our approach enables more sophisticated modeling of such environments, significantly improving the system's ability to adapt to real-world complexities. By combining this method with advanced RL techniques, such as Proximal Policy Optimization (PPO) and Model-Agnostic Meta-Learning (MAML), and incorporating self-learning driven by stability, we improve the system's generalization capabilities. This facilitates rapid learning from novel and previously unseen scenarios. We validate our algorithm through both simulations and real-world experiments on the HRP-4 humanoid robot, utilizing an intrinsically stable model predictive controller.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 4","pages":"3747-3754"},"PeriodicalIF":4.6000,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10904273/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Recent advancements in reinforcement learning (RL) and humanoid robotics are rapidly addressing the challenge of adapting to complex, dynamic environments in real time. This letter introduces a novel approach that integrates two key concepts: experience variability (a criterion for detecting changes in loco-manipulation) and experience accumulation (an efficient method for storing acquired experiences based on a selection criterion). These elements are incorporated into the development of RL agents and humanoid robots, with an emphasis on stability. This focus enhances adaptability and efficiency in unpredictable environments. Our approach enables more sophisticated modeling of such environments, significantly improving the system's ability to adapt to real-world complexities. By combining this method with advanced RL techniques, such as Proximal Policy Optimization (PPO) and Model-Agnostic Meta-Learning (MAML), and incorporating self-learning driven by stability, we improve the system's generalization capabilities. This facilitates rapid learning from novel and previously unseen scenarios. We validate our algorithm through both simulations and real-world experiments on the HRP-4 humanoid robot, utilizing an intrinsically stable model predictive controller.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.