{"title":"面向任务的机器人自主技能学习","authors":"Chenyang Ran, Jianbo Su","doi":"10.1142/s0219843624500014","DOIUrl":null,"url":null,"abstract":"<p>The inferior sample efficiency of reinforcement learning (RL) and the requirement for high-quality demonstrations in imitation learning (IL) will hinder their application in real-world robots. To address this challenge, a novel self-evolution framework, named task-oriented self-imitation learning (TOSIL), is proposed. To circumvent external demonstrations, the top-K self-generated trajectories are chosen as expert data from both per-episode exploration and long-term return perspectives. Each transition is assigned a guide reward, which is formulated by these trajectories. The guide rewards update as the agent evolves, encouraging good exploration behaviors. This methodology guarantees that the agent explores in the direction relevant to the task, improving sample efficiency and asymptotic performance. The experimental results on locomotion and manipulation tasks indicate that the proposed framework outperforms other state-of-the-art RL methods. Furthermore, the integration of suboptimal trajectories has the potential to improve the sample efficiency while maintaining performance. This is a significant advancement in autonomous skill acquisition for robots.</p>","PeriodicalId":50319,"journal":{"name":"International Journal of Humanoid Robotics","volume":"21 1","pages":""},"PeriodicalIF":0.9000,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Task-Oriented Self-Imitation Learning for Robotic Autonomous Skill Acquisition\",\"authors\":\"Chenyang Ran, Jianbo Su\",\"doi\":\"10.1142/s0219843624500014\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>The inferior sample efficiency of reinforcement learning (RL) and the requirement for high-quality demonstrations in imitation learning (IL) will hinder their application in real-world robots. To address this challenge, a novel self-evolution framework, named task-oriented self-imitation learning (TOSIL), is proposed. To circumvent external demonstrations, the top-K self-generated trajectories are chosen as expert data from both per-episode exploration and long-term return perspectives. Each transition is assigned a guide reward, which is formulated by these trajectories. The guide rewards update as the agent evolves, encouraging good exploration behaviors. This methodology guarantees that the agent explores in the direction relevant to the task, improving sample efficiency and asymptotic performance. The experimental results on locomotion and manipulation tasks indicate that the proposed framework outperforms other state-of-the-art RL methods. Furthermore, the integration of suboptimal trajectories has the potential to improve the sample efficiency while maintaining performance. This is a significant advancement in autonomous skill acquisition for robots.</p>\",\"PeriodicalId\":50319,\"journal\":{\"name\":\"International Journal of Humanoid Robotics\",\"volume\":\"21 1\",\"pages\":\"\"},\"PeriodicalIF\":0.9000,\"publicationDate\":\"2024-03-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Humanoid Robotics\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1142/s0219843624500014\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"ROBOTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Humanoid Robotics","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1142/s0219843624500014","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
摘要
强化学习(RL)的样本效率较低,而模仿学习(IL)需要高质量的示范,这将阻碍它们在现实世界机器人中的应用。为了应对这一挑战,我们提出了一个新颖的自我进化框架,名为面向任务的自我模仿学习(TOSIL)。为避免外部示范,从每集探索和长期回报的角度出发,选择前 K 个自生成轨迹作为专家数据。每个过渡都有一个指导奖励,由这些轨迹制定。指导奖励随着代理的发展而更新,以鼓励良好的探索行为。这种方法保证了机器人在与任务相关的方向上进行探索,提高了样本效率和渐进性能。运动和操纵任务的实验结果表明,所提出的框架优于其他最先进的 RL 方法。此外,整合次优轨迹有可能在保持性能的同时提高采样效率。这是机器人自主技能获取领域的一大进步。
Task-Oriented Self-Imitation Learning for Robotic Autonomous Skill Acquisition
The inferior sample efficiency of reinforcement learning (RL) and the requirement for high-quality demonstrations in imitation learning (IL) will hinder their application in real-world robots. To address this challenge, a novel self-evolution framework, named task-oriented self-imitation learning (TOSIL), is proposed. To circumvent external demonstrations, the top-K self-generated trajectories are chosen as expert data from both per-episode exploration and long-term return perspectives. Each transition is assigned a guide reward, which is formulated by these trajectories. The guide rewards update as the agent evolves, encouraging good exploration behaviors. This methodology guarantees that the agent explores in the direction relevant to the task, improving sample efficiency and asymptotic performance. The experimental results on locomotion and manipulation tasks indicate that the proposed framework outperforms other state-of-the-art RL methods. Furthermore, the integration of suboptimal trajectories has the potential to improve the sample efficiency while maintaining performance. This is a significant advancement in autonomous skill acquisition for robots.
期刊介绍:
The International Journal of Humanoid Robotics (IJHR) covers all subjects on the mind and body of humanoid robots. It is dedicated to advancing new theories, new techniques, and new implementations contributing to the successful achievement of future robots which not only imitate human beings, but also serve human beings. While IJHR encourages the contribution of original papers which are solidly grounded on proven theories or experimental procedures, the journal also encourages the contribution of innovative papers which venture into the new, frontier areas in robotics. Such papers need not necessarily demonstrate, in the early stages of research and development, the full potential of new findings on a physical or virtual robot.
IJHR welcomes original papers in the following categories:
Research papers, which disseminate scientific findings contributing to solving technical issues underlying the development of humanoid robots, or biologically-inspired robots, having multiple functionality related to either physical capabilities (i.e. motion) or mental capabilities (i.e. intelligence)
Review articles, which describe, in non-technical terms, the latest in basic theories, principles, and algorithmic solutions
Short articles (e.g. feature articles and dialogues), which discuss the latest significant achievements and the future trends in robotics R&D
Papers on curriculum development in humanoid robot education
Book reviews.