{"title":"基于逆最优控制的离散线性系统逆强化学习。","authors":"Jiashun Huang, Dengguo Xu, Yahui Li, Xiang Zhang, Jingling Zhao","doi":"10.1016/j.isatra.2025.04.027","DOIUrl":null,"url":null,"abstract":"<p><p>This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.</p>","PeriodicalId":94059,"journal":{"name":"ISA transactions","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Inverse reinforcement learning for discrete-time linear systems based on inverse optimal control.\",\"authors\":\"Jiashun Huang, Dengguo Xu, Yahui Li, Xiang Zhang, Jingling Zhao\",\"doi\":\"10.1016/j.isatra.2025.04.027\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.</p>\",\"PeriodicalId\":94059,\"journal\":{\"name\":\"ISA transactions\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-05-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISA transactions\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1016/j.isatra.2025.04.027\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISA transactions","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1016/j.isatra.2025.04.027","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Inverse reinforcement learning for discrete-time linear systems based on inverse optimal control.
This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.