{"title":"一种用于动态目标跟踪的无模型Actor-Critic强化学习方法","authors":"Amr Elhussein, Md. Suruz Miah","doi":"10.1109/MIC50194.2020.9209618","DOIUrl":null,"url":null,"abstract":"Addressing the trajectory tracking problem of a mobile robot in tracking a dynamic target is still one of the challenging problems in the field of robotics. In this paper, we address the position tracking problem of a mobile robot where it is supposed to track the position of a mobile target whose dynamics is unknown a priori. This problem is even more challenging when the dynamics of the mobile robot is also assumed to be unknown, which is indeed a realistic assumption. Most of the trajectory tracking solutions proposed in the literature in the field of mobile robotics are either focused on algorithms that rely on mathematical models of the robots or driven by a overwhelming degree of computational complexity. This paper proposes a model-free actor-critic reinforcement learning strategy to determine appropriate actuator commands for the robot to track the position of the target. We emphasize that mathematical models of both mobile robot and the target are not required in the current approach. Moreover, Bellman’s principle of optimality is utilized to minimize the energy required for the robot to track the target. The performance of the proposed actor-critic reinforcement learning approach is backed by a set of computer experiments with various complexities using a virtual circular-shaped mobile robot and a point target modeled by an integrator.","PeriodicalId":351221,"journal":{"name":"2020 IEEE Midwest Industry Conference (MIC)","volume":"300 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"A Novel Model-Free Actor-Critic Reinforcement Learning Approach for Dynamic Target Tracking\",\"authors\":\"Amr Elhussein, Md. Suruz Miah\",\"doi\":\"10.1109/MIC50194.2020.9209618\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Addressing the trajectory tracking problem of a mobile robot in tracking a dynamic target is still one of the challenging problems in the field of robotics. In this paper, we address the position tracking problem of a mobile robot where it is supposed to track the position of a mobile target whose dynamics is unknown a priori. This problem is even more challenging when the dynamics of the mobile robot is also assumed to be unknown, which is indeed a realistic assumption. Most of the trajectory tracking solutions proposed in the literature in the field of mobile robotics are either focused on algorithms that rely on mathematical models of the robots or driven by a overwhelming degree of computational complexity. This paper proposes a model-free actor-critic reinforcement learning strategy to determine appropriate actuator commands for the robot to track the position of the target. We emphasize that mathematical models of both mobile robot and the target are not required in the current approach. Moreover, Bellman’s principle of optimality is utilized to minimize the energy required for the robot to track the target. The performance of the proposed actor-critic reinforcement learning approach is backed by a set of computer experiments with various complexities using a virtual circular-shaped mobile robot and a point target modeled by an integrator.\",\"PeriodicalId\":351221,\"journal\":{\"name\":\"2020 IEEE Midwest Industry Conference (MIC)\",\"volume\":\"300 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE Midwest Industry Conference (MIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MIC50194.2020.9209618\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Midwest Industry Conference (MIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MIC50194.2020.9209618","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Novel Model-Free Actor-Critic Reinforcement Learning Approach for Dynamic Target Tracking
Addressing the trajectory tracking problem of a mobile robot in tracking a dynamic target is still one of the challenging problems in the field of robotics. In this paper, we address the position tracking problem of a mobile robot where it is supposed to track the position of a mobile target whose dynamics is unknown a priori. This problem is even more challenging when the dynamics of the mobile robot is also assumed to be unknown, which is indeed a realistic assumption. Most of the trajectory tracking solutions proposed in the literature in the field of mobile robotics are either focused on algorithms that rely on mathematical models of the robots or driven by a overwhelming degree of computational complexity. This paper proposes a model-free actor-critic reinforcement learning strategy to determine appropriate actuator commands for the robot to track the position of the target. We emphasize that mathematical models of both mobile robot and the target are not required in the current approach. Moreover, Bellman’s principle of optimality is utilized to minimize the energy required for the robot to track the target. The performance of the proposed actor-critic reinforcement learning approach is backed by a set of computer experiments with various complexities using a virtual circular-shaped mobile robot and a point target modeled by an integrator.