基于深度q -网络的旋转倒立摆系统及其在EdgeX平台上的监测

Ju-Bong Kim, Do-Hyung Kwon, Yong-Geun Hong, Hyun-kyo Lim, Min Suk Kim, Youn-Hee Han
{"title":"基于深度q -网络的旋转倒立摆系统及其在EdgeX平台上的监测","authors":"Ju-Bong Kim, Do-Hyung Kwon, Yong-Geun Hong, Hyun-kyo Lim, Min Suk Kim, Youn-Hee Han","doi":"10.1109/ICAIIC.2019.8668979","DOIUrl":null,"url":null,"abstract":"A rotary inverted pendulum is an unstable and highly nonlinear device and is used as a common model for engineering applications in linear and nonlinear control. In this study, we created a cyber physical system (CPS) to demonstrate that a deep reinforcement learning agent using a rotary inverted pendulum can successfully control a remotely located physical device. The device we created is composed of a cyber environment and physical environment using the Message Queuing Telemetry Transport (MQTT) protocol with an Ethernet connection to connect the cyber environment and the physical environment. The reinforcement learning agent controls the physical device, which is located remotely from the controller and a classical proportional integral derivative (PID) controller is utilized to implement imitation and reinforcement learning and facilitate the learning process. In addition, the control and monitoring system is built on the open source EdgeX platform, so that learning tasks performed near the source of data generation and real-time data emitted from the physical device can be observed while reinforcement learning is performed. From our CPS experimental system, we verify that a deep reinforcement learning agent can control a remotely located real-world device successfully.","PeriodicalId":273383,"journal":{"name":"2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Deep Q-Network Based Rotary Inverted Pendulum System and Its Monitoring on the EdgeX Platform\",\"authors\":\"Ju-Bong Kim, Do-Hyung Kwon, Yong-Geun Hong, Hyun-kyo Lim, Min Suk Kim, Youn-Hee Han\",\"doi\":\"10.1109/ICAIIC.2019.8668979\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A rotary inverted pendulum is an unstable and highly nonlinear device and is used as a common model for engineering applications in linear and nonlinear control. In this study, we created a cyber physical system (CPS) to demonstrate that a deep reinforcement learning agent using a rotary inverted pendulum can successfully control a remotely located physical device. The device we created is composed of a cyber environment and physical environment using the Message Queuing Telemetry Transport (MQTT) protocol with an Ethernet connection to connect the cyber environment and the physical environment. The reinforcement learning agent controls the physical device, which is located remotely from the controller and a classical proportional integral derivative (PID) controller is utilized to implement imitation and reinforcement learning and facilitate the learning process. In addition, the control and monitoring system is built on the open source EdgeX platform, so that learning tasks performed near the source of data generation and real-time data emitted from the physical device can be observed while reinforcement learning is performed. From our CPS experimental system, we verify that a deep reinforcement learning agent can control a remotely located real-world device successfully.\",\"PeriodicalId\":273383,\"journal\":{\"name\":\"2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAIIC.2019.8668979\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAIIC.2019.8668979","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

摘要

旋转式倒立摆是一种不稳定的、高度非线性的装置,是工程应用中线性和非线性控制的常用模型。在本研究中,我们创建了一个网络物理系统(CPS)来证明使用旋转倒立摆的深度强化学习代理可以成功地控制远程物理设备。我们创建的设备由网络环境和物理环境组成,使用消息队列遥测传输(MQTT)协议,并使用以太网连接连接网络环境和物理环境。强化学习代理控制远离控制器的物理设备,利用经典的比例积分导数(PID)控制器实现模仿和强化学习,方便学习过程。此外,控制和监控系统建立在开源的EdgeX平台上,在进行强化学习的同时,可以观察到在数据生成源附近执行的学习任务和从物理设备发出的实时数据。从我们的CPS实验系统中,我们验证了深度强化学习代理可以成功地控制远程位于现实世界的设备。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Deep Q-Network Based Rotary Inverted Pendulum System and Its Monitoring on the EdgeX Platform
A rotary inverted pendulum is an unstable and highly nonlinear device and is used as a common model for engineering applications in linear and nonlinear control. In this study, we created a cyber physical system (CPS) to demonstrate that a deep reinforcement learning agent using a rotary inverted pendulum can successfully control a remotely located physical device. The device we created is composed of a cyber environment and physical environment using the Message Queuing Telemetry Transport (MQTT) protocol with an Ethernet connection to connect the cyber environment and the physical environment. The reinforcement learning agent controls the physical device, which is located remotely from the controller and a classical proportional integral derivative (PID) controller is utilized to implement imitation and reinforcement learning and facilitate the learning process. In addition, the control and monitoring system is built on the open source EdgeX platform, so that learning tasks performed near the source of data generation and real-time data emitted from the physical device can be observed while reinforcement learning is performed. From our CPS experimental system, we verify that a deep reinforcement learning agent can control a remotely located real-world device successfully.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信