{"title":"Comparison of control methods: Learning robotics manipulation with contact dynamics","authors":"Kedao Wang, Yong Li","doi":"10.1109/SIMPAR.2018.8376263","DOIUrl":null,"url":null,"abstract":"We compare the different control methods in learning a robotic manipulation task. The task is to push an object (a cube and sphere) from varying beginning position to a fixed goal position. Complex contact dynamics is involved. We used PPO as the learning algorithm trained from scratch with dense rewards. Comparison is performed on two dimensions: learning at joint level vs. end-effector level, as well as velocity control vs. position control. For end-effector learning, we use inverse jacobian to map from end-effector target velocity/position to joint velocity/position, and accounting for singularity, joint limits, and gimbal lock. Across the four methods proposed, joint velocity control demonstrated the fastest convergence on cube task across all control methods, and is the only successful method on sphere task. Video demonstration: https://www.youtube.com/watch?v=wh_qV58f95Y.","PeriodicalId":156498,"journal":{"name":"2018 IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SIMPAR.2018.8376263","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
We compare the different control methods in learning a robotic manipulation task. The task is to push an object (a cube and sphere) from varying beginning position to a fixed goal position. Complex contact dynamics is involved. We used PPO as the learning algorithm trained from scratch with dense rewards. Comparison is performed on two dimensions: learning at joint level vs. end-effector level, as well as velocity control vs. position control. For end-effector learning, we use inverse jacobian to map from end-effector target velocity/position to joint velocity/position, and accounting for singularity, joint limits, and gimbal lock. Across the four methods proposed, joint velocity control demonstrated the fastest convergence on cube task across all control methods, and is the only successful method on sphere task. Video demonstration: https://www.youtube.com/watch?v=wh_qV58f95Y.