O. Petrovic, Lukas Schäper, Simon Roggendorf, S. Storms, C. Brecher
{"title":"Sim2Real Deep Reinforcement Learning of Compliance-based Robotic Assembly Operations","authors":"O. Petrovic, Lukas Schäper, Simon Roggendorf, S. Storms, C. Brecher","doi":"10.1109/MMAR55195.2022.9874304","DOIUrl":null,"url":null,"abstract":"Reinforcement learning (RL) enables robots to learn goal-oriented behavior. In production processes with high variances, such as joining operations in end-of-line assembly, this is particularly interesting to save significant programming effort. Due to a large amount of required training data, simulative training is becoming increasingly important. In this paper, we present an approach to learn a contact-rich peg-in-hole assembly task utilizing deep reinforcement learning (DRL) and a compliant robot controller. The DRL-Agent learns directly in the Cartesian space (task space) and not in the joint space of the robot, to increase the robustness and efficiency of the algorithms. To further increase the robustness of the policy and to shorten training times, geometric limitations are imposed by introducing an admissible workspace using a trajectory generator. Furthermore, these limitations result in nearly identical behavior in the simulation and on the real robot, allowing the DRL training process to be purely simulative. The learned policy is experimentally investigated both in the simulation environment and on a real robot, to evaluate its transferability from simulation to reality (sim2real).","PeriodicalId":169528,"journal":{"name":"2022 26th International Conference on Methods and Models in Automation and Robotics (MMAR)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 26th International Conference on Methods and Models in Automation and Robotics (MMAR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MMAR55195.2022.9874304","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Reinforcement learning (RL) enables robots to learn goal-oriented behavior. In production processes with high variances, such as joining operations in end-of-line assembly, this is particularly interesting to save significant programming effort. Due to a large amount of required training data, simulative training is becoming increasingly important. In this paper, we present an approach to learn a contact-rich peg-in-hole assembly task utilizing deep reinforcement learning (DRL) and a compliant robot controller. The DRL-Agent learns directly in the Cartesian space (task space) and not in the joint space of the robot, to increase the robustness and efficiency of the algorithms. To further increase the robustness of the policy and to shorten training times, geometric limitations are imposed by introducing an admissible workspace using a trajectory generator. Furthermore, these limitations result in nearly identical behavior in the simulation and on the real robot, allowing the DRL training process to be purely simulative. The learned policy is experimentally investigated both in the simulation environment and on a real robot, to evaluate its transferability from simulation to reality (sim2real).