一种用于动态目标跟踪的无模型Actor-Critic强化学习方法

Amr Elhussein, Md. Suruz Miah
{"title":"一种用于动态目标跟踪的无模型Actor-Critic强化学习方法","authors":"Amr Elhussein, Md. Suruz Miah","doi":"10.1109/MIC50194.2020.9209618","DOIUrl":null,"url":null,"abstract":"Addressing the trajectory tracking problem of a mobile robot in tracking a dynamic target is still one of the challenging problems in the field of robotics. In this paper, we address the position tracking problem of a mobile robot where it is supposed to track the position of a mobile target whose dynamics is unknown a priori. This problem is even more challenging when the dynamics of the mobile robot is also assumed to be unknown, which is indeed a realistic assumption. Most of the trajectory tracking solutions proposed in the literature in the field of mobile robotics are either focused on algorithms that rely on mathematical models of the robots or driven by a overwhelming degree of computational complexity. This paper proposes a model-free actor-critic reinforcement learning strategy to determine appropriate actuator commands for the robot to track the position of the target. We emphasize that mathematical models of both mobile robot and the target are not required in the current approach. Moreover, Bellman’s principle of optimality is utilized to minimize the energy required for the robot to track the target. The performance of the proposed actor-critic reinforcement learning approach is backed by a set of computer experiments with various complexities using a virtual circular-shaped mobile robot and a point target modeled by an integrator.","PeriodicalId":351221,"journal":{"name":"2020 IEEE Midwest Industry Conference (MIC)","volume":"300 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"A Novel Model-Free Actor-Critic Reinforcement Learning Approach for Dynamic Target Tracking\",\"authors\":\"Amr Elhussein, Md. Suruz Miah\",\"doi\":\"10.1109/MIC50194.2020.9209618\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Addressing the trajectory tracking problem of a mobile robot in tracking a dynamic target is still one of the challenging problems in the field of robotics. In this paper, we address the position tracking problem of a mobile robot where it is supposed to track the position of a mobile target whose dynamics is unknown a priori. This problem is even more challenging when the dynamics of the mobile robot is also assumed to be unknown, which is indeed a realistic assumption. Most of the trajectory tracking solutions proposed in the literature in the field of mobile robotics are either focused on algorithms that rely on mathematical models of the robots or driven by a overwhelming degree of computational complexity. This paper proposes a model-free actor-critic reinforcement learning strategy to determine appropriate actuator commands for the robot to track the position of the target. We emphasize that mathematical models of both mobile robot and the target are not required in the current approach. Moreover, Bellman’s principle of optimality is utilized to minimize the energy required for the robot to track the target. The performance of the proposed actor-critic reinforcement learning approach is backed by a set of computer experiments with various complexities using a virtual circular-shaped mobile robot and a point target modeled by an integrator.\",\"PeriodicalId\":351221,\"journal\":{\"name\":\"2020 IEEE Midwest Industry Conference (MIC)\",\"volume\":\"300 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE Midwest Industry Conference (MIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MIC50194.2020.9209618\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Midwest Industry Conference (MIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MIC50194.2020.9209618","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

摘要

如何解决移动机器人在跟踪动态目标时的轨迹跟踪问题,一直是机器人领域的难题之一。在本文中,我们解决了一个移动机器人的位置跟踪问题,它应该跟踪一个动态未知的先验移动目标的位置。当移动机器人的动力学也被假设为未知时,这个问题更具挑战性,这确实是一个现实的假设。在移动机器人领域的文献中提出的大多数轨迹跟踪解决方案要么集中在依赖于机器人数学模型的算法上,要么由压倒性的计算复杂性驱动。本文提出了一种无模型行为者评价强化学习策略,以确定机器人跟踪目标位置的适当执行器命令。我们强调,在目前的方法中,不需要同时建立移动机器人和目标的数学模型。此外,利用Bellman的最优性原理使机器人跟踪目标所需的能量最小。所提出的行为者-评论家强化学习方法的性能得到了一系列具有各种复杂性的计算机实验的支持,这些实验使用了一个虚拟圆形移动机器人和一个由积分器建模的点目标。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Novel Model-Free Actor-Critic Reinforcement Learning Approach for Dynamic Target Tracking
Addressing the trajectory tracking problem of a mobile robot in tracking a dynamic target is still one of the challenging problems in the field of robotics. In this paper, we address the position tracking problem of a mobile robot where it is supposed to track the position of a mobile target whose dynamics is unknown a priori. This problem is even more challenging when the dynamics of the mobile robot is also assumed to be unknown, which is indeed a realistic assumption. Most of the trajectory tracking solutions proposed in the literature in the field of mobile robotics are either focused on algorithms that rely on mathematical models of the robots or driven by a overwhelming degree of computational complexity. This paper proposes a model-free actor-critic reinforcement learning strategy to determine appropriate actuator commands for the robot to track the position of the target. We emphasize that mathematical models of both mobile robot and the target are not required in the current approach. Moreover, Bellman’s principle of optimality is utilized to minimize the energy required for the robot to track the target. The performance of the proposed actor-critic reinforcement learning approach is backed by a set of computer experiments with various complexities using a virtual circular-shaped mobile robot and a point target modeled by an integrator.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信