M. Shetty, Brunda Vishishta, Shrinidhi Choragi, Karpagavalli Subramanian, Koshy George
{"title":"Continuous Control of a Robot Manipulator Using Deep Deterministic Policy Gradient","authors":"M. Shetty, Brunda Vishishta, Shrinidhi Choragi, Karpagavalli Subramanian, Koshy George","doi":"10.1109/ICC54714.2021.9703155","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning (DRL) addresses the problems that previously limited the performance of RL algorithms while working with high-dimensional state and action spaces. In this paper, we explore the deep deterministic policy gradient (DDPG) algorithm that operates over continuous action spaces. The application of reference tracking for a two-link robot manipulator (TLRM) in uncertain environments is considered. The TLRM is subjected to uncertainties such as frictional forces and external torque disturbances. In the simulation study, we compare the performance of our RL-based controller with the well-known proportional-derivative (PD) controller. Results indicate a considerable improvement in the mean square error (MSE) and variance accounted for (VAF) metrics when the RL-based controller is utilized.","PeriodicalId":382373,"journal":{"name":"2021 Seventh Indian Control Conference (ICC)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 Seventh Indian Control Conference (ICC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICC54714.2021.9703155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Deep reinforcement learning (DRL) addresses the problems that previously limited the performance of RL algorithms while working with high-dimensional state and action spaces. In this paper, we explore the deep deterministic policy gradient (DDPG) algorithm that operates over continuous action spaces. The application of reference tracking for a two-link robot manipulator (TLRM) in uncertain environments is considered. The TLRM is subjected to uncertainties such as frictional forces and external torque disturbances. In the simulation study, we compare the performance of our RL-based controller with the well-known proportional-derivative (PD) controller. Results indicate a considerable improvement in the mean square error (MSE) and variance accounted for (VAF) metrics when the RL-based controller is utilized.