Reduction of Complexity in Q-Learning a Robot Control for an Assembly Cell by using Multiple Agents

Georg Kunert, T. Pawletta, Sven Hartmann
{"title":"Reduction of Complexity in Q-Learning a Robot Control for an Assembly Cell by using Multiple Agents","authors":"Georg Kunert, T. Pawletta, Sven Hartmann","doi":"10.11128/sne.30.tn.10524","DOIUrl":null,"url":null,"abstract":"Production systems in Industry 4.0 are characterized by a high degree of system networking and adaptability. They are often characterized by jointed-arm robots, which have a high degree of adaptation. Networking and adaptivity increase the flexibility of a system, but also the complexity of the control, which requires the use of new development methods. In this context, the Simulation-Based Control approach, a model-based design method, and the concept of Reinforcement Learning (RL) are introduced and it is shown how a task-based robot control can be learned and executed. Afterwards, the time complexity of the Q-learning method will be examined using the application example of a robot-based assembly cell with two differently flexible system configurations. It is shown that, depending on the system configuration, the time complexity of learning can be significantly reduced when using several agents. In the studied case, the complexity decreases from exponential to linear. The modified RL structure is discussed in detail.","PeriodicalId":262785,"journal":{"name":"Simul. Notes Eur.","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Simul. Notes Eur.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.11128/sne.30.tn.10524","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Production systems in Industry 4.0 are characterized by a high degree of system networking and adaptability. They are often characterized by jointed-arm robots, which have a high degree of adaptation. Networking and adaptivity increase the flexibility of a system, but also the complexity of the control, which requires the use of new development methods. In this context, the Simulation-Based Control approach, a model-based design method, and the concept of Reinforcement Learning (RL) are introduced and it is shown how a task-based robot control can be learned and executed. Afterwards, the time complexity of the Q-learning method will be examined using the application example of a robot-based assembly cell with two differently flexible system configurations. It is shown that, depending on the system configuration, the time complexity of learning can be significantly reduced when using several agents. In the studied case, the complexity decreases from exponential to linear. The modified RL structure is discussed in detail.
基于多智能体的机器人装配单元控制q -学习复杂度降低
工业4.0的生产系统的特点是高度的系统联网和适应性。它们的特点往往是关节臂机器人,具有高度的适应性。网络化和自适应增加了系统的灵活性,但也增加了控制的复杂性,这就需要使用新的开发方法。在此背景下,介绍了基于仿真的控制方法、基于模型的设计方法和强化学习(RL)的概念,并展示了如何学习和执行基于任务的机器人控制。随后,将使用具有两种不同柔性系统配置的机器人装配单元的应用实例来检验q -学习方法的时间复杂度。研究表明,根据不同的系统配置,当使用多个智能体时,学习的时间复杂度可以显著降低。在研究的情况下,复杂度从指数下降到线性。详细讨论了改进后的RL结构。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信