Inverse reinforcement learning for discrete-time linear systems based on inverse optimal control.

Jiashun Huang, Dengguo Xu, Yahui Li, Xiang Zhang, Jingling Zhao
{"title":"Inverse reinforcement learning for discrete-time linear systems based on inverse optimal control.","authors":"Jiashun Huang, Dengguo Xu, Yahui Li, Xiang Zhang, Jingling Zhao","doi":"10.1016/j.isatra.2025.04.027","DOIUrl":null,"url":null,"abstract":"<p><p>This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.</p>","PeriodicalId":94059,"journal":{"name":"ISA transactions","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISA transactions","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1016/j.isatra.2025.04.027","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

This paper mainly deals with inverse reinforcement learning (IRL) for discrete-time linear time-invariant systems. Based on input and state measurement data from expert agent, several algorithms are proposed to reconstruct cost function in optimal control problem. The algorithms mainly consist of three steps, namely updating control gain via algebraic Riccati equation (ARE), gradient descent to correct cost matrix, and updating weight matrix based on inverse optimal control (IOC). First, by reformulating gain formula of optimal control in the learner system, we present a model-based IRL algorithm. When the system model is fully known, the cost function can be iteratively computed. Then, we develop a partially model-free IRL framework for reconstructing the cost function by introducing auxiliary control inputs and decomposing the algorithm into outer and inner loop. Therefore, in the case where the input matrix is unknown, weight matrix in the cost function is reconstructed. Moreover, the convergence of the algorithms and the stability of corresponding closed-loop system have been demonstrated. Finally, simulations verify the effectiveness of the proposed IRL algorithms.

基于逆最优控制的离散线性系统逆强化学习。
本文主要研究离散线性定常系统的逆强化学习(IRL)。基于专家智能体的输入和状态测量数据,提出了几种重构最优控制问题中代价函数的算法。该算法主要包括三个步骤,即通过代数Riccati方程(ARE)更新控制增益、梯度下降修正代价矩阵和基于逆最优控制(IOC)更新权矩阵。首先,通过重新表述学习系统中最优控制的增益公式,提出了一种基于模型的IRL算法。当系统模型完全已知时,可以迭代计算成本函数。然后,我们开发了一个部分无模型的IRL框架,通过引入辅助控制输入并将算法分解为外部和内部循环来重建成本函数。因此,在输入矩阵未知的情况下,重构代价函数中的权矩阵。此外,还证明了算法的收敛性和相应闭环系统的稳定性。最后,通过仿真验证了所提IRL算法的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信