{"title":"基于初始激励的连续时间线性非零和博弈逆强化学习","authors":"Hongyang Li, Gansu Zhang, Qinglai Wei","doi":"10.1002/rnc.8026","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>In this article, the initial excitation-based inverse reinforcement learning methods are presented for continuous-time linear non-zero-sum games. The policy iteration and value iteration algorithms are presented for the inverse reinforcement learning problems, and an online-verifiable initial excitation condition is given to guarantee the convergence of the presented algorithms. Comparing with the traditional inverse reinforcement learning algorithms for linear non-zero-sum games, the presented algorithms relax the requirement on data-storage mechanism. Furthermore, the requirement on the initial stabilizing state feedback matrices is relaxed in the presented initial excitation-based value iteration algorithm. The properties of the presented initial excitation-based policy iteration and value iteration algorithms are analyzed. Simulation results show the efficiency of the presented algorithms.</p>\n </div>","PeriodicalId":50291,"journal":{"name":"International Journal of Robust and Nonlinear Control","volume":"35 15","pages":"6275-6287"},"PeriodicalIF":3.2000,"publicationDate":"2025-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Initial Excitation-Based Inverse Reinforcement Learning for Continuous-Time Linear Non-Zero-Sum Games\",\"authors\":\"Hongyang Li, Gansu Zhang, Qinglai Wei\",\"doi\":\"10.1002/rnc.8026\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <p>In this article, the initial excitation-based inverse reinforcement learning methods are presented for continuous-time linear non-zero-sum games. The policy iteration and value iteration algorithms are presented for the inverse reinforcement learning problems, and an online-verifiable initial excitation condition is given to guarantee the convergence of the presented algorithms. Comparing with the traditional inverse reinforcement learning algorithms for linear non-zero-sum games, the presented algorithms relax the requirement on data-storage mechanism. Furthermore, the requirement on the initial stabilizing state feedback matrices is relaxed in the presented initial excitation-based value iteration algorithm. The properties of the presented initial excitation-based policy iteration and value iteration algorithms are analyzed. Simulation results show the efficiency of the presented algorithms.</p>\\n </div>\",\"PeriodicalId\":50291,\"journal\":{\"name\":\"International Journal of Robust and Nonlinear Control\",\"volume\":\"35 15\",\"pages\":\"6275-6287\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-05-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Robust and Nonlinear Control\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/rnc.8026\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Robust and Nonlinear Control","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/rnc.8026","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Initial Excitation-Based Inverse Reinforcement Learning for Continuous-Time Linear Non-Zero-Sum Games
In this article, the initial excitation-based inverse reinforcement learning methods are presented for continuous-time linear non-zero-sum games. The policy iteration and value iteration algorithms are presented for the inverse reinforcement learning problems, and an online-verifiable initial excitation condition is given to guarantee the convergence of the presented algorithms. Comparing with the traditional inverse reinforcement learning algorithms for linear non-zero-sum games, the presented algorithms relax the requirement on data-storage mechanism. Furthermore, the requirement on the initial stabilizing state feedback matrices is relaxed in the presented initial excitation-based value iteration algorithm. The properties of the presented initial excitation-based policy iteration and value iteration algorithms are analyzed. Simulation results show the efficiency of the presented algorithms.
期刊介绍:
Papers that do not include an element of robust or nonlinear control and estimation theory will not be considered by the journal, and all papers will be expected to include significant novel content. The focus of the journal is on model based control design approaches rather than heuristic or rule based methods. Papers on neural networks will have to be of exceptional novelty to be considered for the journal.