{"title":"Optimal Cooperative Control of Multi-Agent Systems Through Event-Triggered Model-Free Reinforcement Learning","authors":"Chaoxu Mu;Zhuo Tang;Ke Wang","doi":"10.1109/TETCI.2024.3451484","DOIUrl":null,"url":null,"abstract":"This paper addresses the optimal cooperative control problem for nonlinear multi-agent systems with completely unknown dynamics and proposes a learning control scheme based on the event-triggered mechanisms. The problem is reformulated as a multi-agent differential graphical game, and an off-policy integral reinforcement learning algorithm is introduced by deriving off-policy Bellman equations. To reduce the computational burden of the controller, an event-triggered mechanism is integrated into the adaptive learning process. To overcome the limitations of static triggering, the dynamic variable is introduced to utilize past triggering information. The theoretical proof demonstrates the asymptotic stability of the system and a numerical example validates the effectiveness of the proposed control scheme. Finally, in the case of the multiple manipulator system, a comparison of four control schemes shows that the proposed method not only ensures the system's control performance but also achieves a larger triggering interval, reducing the update frequency of the controller and saving communication bandwidth.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 2","pages":"1699-1711"},"PeriodicalIF":5.3000,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10665747/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
This paper addresses the optimal cooperative control problem for nonlinear multi-agent systems with completely unknown dynamics and proposes a learning control scheme based on the event-triggered mechanisms. The problem is reformulated as a multi-agent differential graphical game, and an off-policy integral reinforcement learning algorithm is introduced by deriving off-policy Bellman equations. To reduce the computational burden of the controller, an event-triggered mechanism is integrated into the adaptive learning process. To overcome the limitations of static triggering, the dynamic variable is introduced to utilize past triggering information. The theoretical proof demonstrates the asymptotic stability of the system and a numerical example validates the effectiveness of the proposed control scheme. Finally, in the case of the multiple manipulator system, a comparison of four control schemes shows that the proposed method not only ensures the system's control performance but also achieves a larger triggering interval, reducing the update frequency of the controller and saving communication bandwidth.
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.