{"title":"基于安全强化学习和协方差矩阵自适应的软体机器人自主控制","authors":"Shaswat Garg , Masoud Goharimanesh , Sina Sajjadi , Farrokh Janabi-Sharifi","doi":"10.1016/j.engappai.2025.110791","DOIUrl":null,"url":null,"abstract":"<div><div>The control of soft robots (such as continuum robots) poses significant challenges due to their coupled dynamics with significant inherent nonlinearities. Recently, model-free reinforcement learning algorithms have been proposed as an attractive alternative to model-based methods to address such a challenging control problem through unsupervised learning. However, the safety of robots is usually ignored while training such algorithms. This is particularly important for medical applications of soft robots. Also, the curse of dimensionality in soft robots makes it difficult for a reinforcement learning algorithm to develop an optimal controller. In this work, we propose a safe phasic soft actor–critic algorithm with a covariance matrix adaptation network which is then tested on different soft robots. We demonstrate that the proposed algorithm could learn an optimal policy quickly while satisfying the safety constraints. We formulated and tested our algorithm for (i) multigait soft robot; (ii) soft gripper robot; and (iii) soft robotic trunk. The proposed algorithm achieved an average of 150% higher rewards compared to other state-of-the-art algorithms. Also, adding the safety layer helped reduce the tracking error by 8 times when compared to the algorithm without a safety layer. The policy is validated in Simulation Open Framework Architecture (SOFA) simulations against other state-of-the-art algorithms in terms of tracking errors.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"153 ","pages":"Article 110791"},"PeriodicalIF":8.0000,"publicationDate":"2025-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Autonomous control of soft robots using safe reinforcement learning and covariance matrix adaptation\",\"authors\":\"Shaswat Garg , Masoud Goharimanesh , Sina Sajjadi , Farrokh Janabi-Sharifi\",\"doi\":\"10.1016/j.engappai.2025.110791\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The control of soft robots (such as continuum robots) poses significant challenges due to their coupled dynamics with significant inherent nonlinearities. Recently, model-free reinforcement learning algorithms have been proposed as an attractive alternative to model-based methods to address such a challenging control problem through unsupervised learning. However, the safety of robots is usually ignored while training such algorithms. This is particularly important for medical applications of soft robots. Also, the curse of dimensionality in soft robots makes it difficult for a reinforcement learning algorithm to develop an optimal controller. In this work, we propose a safe phasic soft actor–critic algorithm with a covariance matrix adaptation network which is then tested on different soft robots. We demonstrate that the proposed algorithm could learn an optimal policy quickly while satisfying the safety constraints. We formulated and tested our algorithm for (i) multigait soft robot; (ii) soft gripper robot; and (iii) soft robotic trunk. The proposed algorithm achieved an average of 150% higher rewards compared to other state-of-the-art algorithms. Also, adding the safety layer helped reduce the tracking error by 8 times when compared to the algorithm without a safety layer. The policy is validated in Simulation Open Framework Architecture (SOFA) simulations against other state-of-the-art algorithms in terms of tracking errors.</div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"153 \",\"pages\":\"Article 110791\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-04-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625007912\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625007912","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Autonomous control of soft robots using safe reinforcement learning and covariance matrix adaptation
The control of soft robots (such as continuum robots) poses significant challenges due to their coupled dynamics with significant inherent nonlinearities. Recently, model-free reinforcement learning algorithms have been proposed as an attractive alternative to model-based methods to address such a challenging control problem through unsupervised learning. However, the safety of robots is usually ignored while training such algorithms. This is particularly important for medical applications of soft robots. Also, the curse of dimensionality in soft robots makes it difficult for a reinforcement learning algorithm to develop an optimal controller. In this work, we propose a safe phasic soft actor–critic algorithm with a covariance matrix adaptation network which is then tested on different soft robots. We demonstrate that the proposed algorithm could learn an optimal policy quickly while satisfying the safety constraints. We formulated and tested our algorithm for (i) multigait soft robot; (ii) soft gripper robot; and (iii) soft robotic trunk. The proposed algorithm achieved an average of 150% higher rewards compared to other state-of-the-art algorithms. Also, adding the safety layer helped reduce the tracking error by 8 times when compared to the algorithm without a safety layer. The policy is validated in Simulation Open Framework Architecture (SOFA) simulations against other state-of-the-art algorithms in terms of tracking errors.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.