{"title":"Reinforcement Learning of Whole-Body Control Strategies to Balance a Dynamically Stable Mobile Manipulator","authors":"Vighnesh Vatsal, B. Purushothaman","doi":"10.1109/ICC54714.2021.9703140","DOIUrl":null,"url":null,"abstract":"Mobile manipulators consist of a ground robot base and a mounted robotic arm, with the two components typically controlled as separate subsystems. This is enabled by the fact that most mobile bases with three or four-wheeled designs are inherently stable, though lacking in maneuverability. In contrast, dynamically stable mobile bases offer greater agility and safety in crowded human interaction scenarios, though requiring active balancing. In this work, we consider the balancing problem for a Two-Wheeled Inverted Pendulum Mobile Manipulator (TWIP-MM), designed for retail shelf inspection. Using deep reinforcement learning methods (PPO and SAC), we can generate whole-body control strategies that leverage the motion of the robotic arm for in-place stabilization of the base, through a completely model-free approach. In contrast, tuning a standard PID controller requires a model of the robot, and is considered here as a baseline. Compared to PID control in simulation, the RL-based controllers are found to be more robust against changes in initial conditions, variations in inertial parameters, and disturbances applied to the robot.","PeriodicalId":382373,"journal":{"name":"2021 Seventh Indian Control Conference (ICC)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 Seventh Indian Control Conference (ICC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICC54714.2021.9703140","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Mobile manipulators consist of a ground robot base and a mounted robotic arm, with the two components typically controlled as separate subsystems. This is enabled by the fact that most mobile bases with three or four-wheeled designs are inherently stable, though lacking in maneuverability. In contrast, dynamically stable mobile bases offer greater agility and safety in crowded human interaction scenarios, though requiring active balancing. In this work, we consider the balancing problem for a Two-Wheeled Inverted Pendulum Mobile Manipulator (TWIP-MM), designed for retail shelf inspection. Using deep reinforcement learning methods (PPO and SAC), we can generate whole-body control strategies that leverage the motion of the robotic arm for in-place stabilization of the base, through a completely model-free approach. In contrast, tuning a standard PID controller requires a model of the robot, and is considered here as a baseline. Compared to PID control in simulation, the RL-based controllers are found to be more robust against changes in initial conditions, variations in inertial parameters, and disturbances applied to the robot.