{"title":"Greedy-DAgger - A Student Rollout Efficient Imitation Learning Algorithm","authors":"Mitchell Torok;Mohammad Deghat;Yang Song","doi":"10.1109/LRA.2025.3536297","DOIUrl":null,"url":null,"abstract":"Sampling-based model predictive control algorithms can be computationally expensive and may not be feasible for restricted platforms such as quadcopters. Comparatively speaking, lightweight learned controllers are computationally cheaper and may be more suited for these platforms. Expert control samples provided by a remote model predictive control algorithm could be used to rapidly train a student policy. We present Greedy-DAgger, a hybrid-policy imitation learning approach that leverages expert simulations to improve the student rollout efficiency during the training of a student policy. Our approach builds on the DAgger algorithm by employing a greedy strategy, that selects isolated states from a student trajectory. These states are used to generate expert trajectory samples before supervised learning is performed and the process is repeated. The effectiveness of the Greedy-DAgger algorithm is evaluated on two simulated robotic systems: a cart pole and a quadcopter. For these environments, Greedy-DAgger was shown to be up to ten times more rollout efficient than conventional DAgger. The introduced improvements enable expert-level quadcopter control to be achieved within 8 seconds of wall time. The Crazyflie quadcopter platform was then utilised to validate the simulation results and demonstrate the potential for real-world training with Greedy-DAgger on a constrained platform, leveraging access to a remote GPU-accelerated server.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 3","pages":"2878-2885"},"PeriodicalIF":4.6000,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10857457/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Sampling-based model predictive control algorithms can be computationally expensive and may not be feasible for restricted platforms such as quadcopters. Comparatively speaking, lightweight learned controllers are computationally cheaper and may be more suited for these platforms. Expert control samples provided by a remote model predictive control algorithm could be used to rapidly train a student policy. We present Greedy-DAgger, a hybrid-policy imitation learning approach that leverages expert simulations to improve the student rollout efficiency during the training of a student policy. Our approach builds on the DAgger algorithm by employing a greedy strategy, that selects isolated states from a student trajectory. These states are used to generate expert trajectory samples before supervised learning is performed and the process is repeated. The effectiveness of the Greedy-DAgger algorithm is evaluated on two simulated robotic systems: a cart pole and a quadcopter. For these environments, Greedy-DAgger was shown to be up to ten times more rollout efficient than conventional DAgger. The introduced improvements enable expert-level quadcopter control to be achieved within 8 seconds of wall time. The Crazyflie quadcopter platform was then utilised to validate the simulation results and demonstrate the potential for real-world training with Greedy-DAgger on a constrained platform, leveraging access to a remote GPU-accelerated server.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.