{"title":"利用深度强化学习学习机器人按摩的可变阻抗控制:新颖的学习框架","authors":"Zhuoran Li, Chao Zeng, Zhen Deng, Qinling Xu, Bingwei He, Jianwei Zhang","doi":"10.1109/MSMC.2022.3231416","DOIUrl":null,"url":null,"abstract":"Contact-rich manipulation tasks are difficult to program to be performed by robots. Traditional compliance control methods, such as impedance control, rely excessively on environmental models and are ineffective in the face of increasingly complex contact tasks. Reinforcement learning (RL) has now achieved great success in the fields of games and robotics. Autonomous learning of manipulation skills can empower robots with autonomous decision-making capabilities. To this end, this work introduces a novel learning framework that combines deep RL (DRL) and variable impedance control (VIC) to achieve robotic massage tasks. A skill policy is learned in joint space, which outputs the desired impedance gain and angle for each joint. To address the limitations of the sparse reward of DRL, an intrinsic curiosity module (ICM) was designed, which generates the intrinsic reward to encourage robots to explore more effectively. Simulation and real experiments were performed to verify the effectiveness of the proposed method. Our experiments demonstrate that contact-rich massage skills can be learned through the VIC–DRL framework based on the joint space in a simulation environment, and that the ICM can improve learning efficiency and overall performance in the task. Moreover, the generated policies have been demonstrated to still perform effectively on a real-world robot.","PeriodicalId":516814,"journal":{"name":"IEEE Systems, Man, and Cybernetics Magazine","volume":"48 4","pages":"17-27"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Learning Variable Impedance Control for Robotic Massage With Deep Reinforcement Learning: A Novel Learning Framework\",\"authors\":\"Zhuoran Li, Chao Zeng, Zhen Deng, Qinling Xu, Bingwei He, Jianwei Zhang\",\"doi\":\"10.1109/MSMC.2022.3231416\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Contact-rich manipulation tasks are difficult to program to be performed by robots. Traditional compliance control methods, such as impedance control, rely excessively on environmental models and are ineffective in the face of increasingly complex contact tasks. Reinforcement learning (RL) has now achieved great success in the fields of games and robotics. Autonomous learning of manipulation skills can empower robots with autonomous decision-making capabilities. To this end, this work introduces a novel learning framework that combines deep RL (DRL) and variable impedance control (VIC) to achieve robotic massage tasks. A skill policy is learned in joint space, which outputs the desired impedance gain and angle for each joint. To address the limitations of the sparse reward of DRL, an intrinsic curiosity module (ICM) was designed, which generates the intrinsic reward to encourage robots to explore more effectively. Simulation and real experiments were performed to verify the effectiveness of the proposed method. Our experiments demonstrate that contact-rich massage skills can be learned through the VIC–DRL framework based on the joint space in a simulation environment, and that the ICM can improve learning efficiency and overall performance in the task. Moreover, the generated policies have been demonstrated to still perform effectively on a real-world robot.\",\"PeriodicalId\":516814,\"journal\":{\"name\":\"IEEE Systems, Man, and Cybernetics Magazine\",\"volume\":\"48 4\",\"pages\":\"17-27\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Systems, Man, and Cybernetics Magazine\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MSMC.2022.3231416\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Systems, Man, and Cybernetics Magazine","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MSMC.2022.3231416","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Learning Variable Impedance Control for Robotic Massage With Deep Reinforcement Learning: A Novel Learning Framework
Contact-rich manipulation tasks are difficult to program to be performed by robots. Traditional compliance control methods, such as impedance control, rely excessively on environmental models and are ineffective in the face of increasingly complex contact tasks. Reinforcement learning (RL) has now achieved great success in the fields of games and robotics. Autonomous learning of manipulation skills can empower robots with autonomous decision-making capabilities. To this end, this work introduces a novel learning framework that combines deep RL (DRL) and variable impedance control (VIC) to achieve robotic massage tasks. A skill policy is learned in joint space, which outputs the desired impedance gain and angle for each joint. To address the limitations of the sparse reward of DRL, an intrinsic curiosity module (ICM) was designed, which generates the intrinsic reward to encourage robots to explore more effectively. Simulation and real experiments were performed to verify the effectiveness of the proposed method. Our experiments demonstrate that contact-rich massage skills can be learned through the VIC–DRL framework based on the joint space in a simulation environment, and that the ICM can improve learning efficiency and overall performance in the task. Moreover, the generated policies have been demonstrated to still perform effectively on a real-world robot.