{"title":"Inverse reinforcement learning for discrete-time linear systems based on inverse optimal control.","authors":"Jiashun Huang, Dengguo Xu, Yahui Li, Xiang Zhang, Jingling Zhao","doi":"10.1016/j.isatra.2025.04.027","DOIUrl":null,"url":null,"abstract":"<p><p>This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.</p>","PeriodicalId":94059,"journal":{"name":"ISA transactions","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISA transactions","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1016/j.isatra.2025.04.027","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.