{"title":"Balance Control of a Humanoid Robot Using DeepReinforcement Learning","authors":"E. Kouchaki, M. Palhang","doi":"10.1109/CSICC58665.2023.10105418","DOIUrl":null,"url":null,"abstract":"In this paper, a deep reinforcement learning algorithm is presented to control a humanoid robot. We have used two control levels in a hierarchical manner. Within the high-level control architecture, a policy is determined by a combination of two neural networks as actor and critic and optimized using proximal policy optimization (PPO) method. The output policy specifies reference angles for robot joint space. At the low-level control, a PID controller regulates robot states around the reference values. The robot model is provided in MuJoCo physics engine and simulations are performed using mujoco-py library. During the simulations robot could maintain its balance stability against wide variety of exerted disturbances. The results showed that the proposed algorithm had a good performance and could resist larger push impacts compared to the pure PID controller.","PeriodicalId":127277,"journal":{"name":"2023 28th International Computer Conference, Computer Society of Iran (CSICC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 28th International Computer Conference, Computer Society of Iran (CSICC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSICC58665.2023.10105418","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, a deep reinforcement learning algorithm is presented to control a humanoid robot. We have used two control levels in a hierarchical manner. Within the high-level control architecture, a policy is determined by a combination of two neural networks as actor and critic and optimized using proximal policy optimization (PPO) method. The output policy specifies reference angles for robot joint space. At the low-level control, a PID controller regulates robot states around the reference values. The robot model is provided in MuJoCo physics engine and simulations are performed using mujoco-py library. During the simulations robot could maintain its balance stability against wide variety of exerted disturbances. The results showed that the proposed algorithm had a good performance and could resist larger push impacts compared to the pure PID controller.