Distributed Nash equilibrium solution for multi-agent game in adversarial environment: A reinforcement learning method

IF 4.8 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS
Qiwei Liu , Huaicheng Yan , Kaitian Chen , Meng Wang , Zhichen Li
{"title":"Distributed Nash equilibrium solution for multi-agent game in adversarial environment: A reinforcement learning method","authors":"Qiwei Liu ,&nbsp;Huaicheng Yan ,&nbsp;Kaitian Chen ,&nbsp;Meng Wang ,&nbsp;Zhichen Li","doi":"10.1016/j.automatica.2025.112342","DOIUrl":null,"url":null,"abstract":"<div><div>This paper investigates the leader–follower optimal consensus problem for linear multi-agent systems with adversarial inputs from a differential graphical game perspective. For the multi-agent optimal control problem described by differential graphical game, it is equally significant that the control policy is distributed and adheres to Nash equilibrium solution. However, achieving both distributed control and Nash equilibrium simultaneously has proven impossible in most existing game formulations. This paper proposes a new game formulation that can overcome this limitation, enabling each agent to reach a Nash equilibrium under a distributed policy, thereby improving system performance. Furthermore, a partially model-free reinforcement learning method is employed to obtain the optimal policy when the dynamic information is partially unknown, with admissibility condition of the initial policy further relaxed. Finally, two comparative simulations are presented to demonstrate the validity and superiority of the proposed approach.</div></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":"178 ","pages":"Article 112342"},"PeriodicalIF":4.8000,"publicationDate":"2025-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Automatica","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0005109825002353","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

This paper investigates the leader–follower optimal consensus problem for linear multi-agent systems with adversarial inputs from a differential graphical game perspective. For the multi-agent optimal control problem described by differential graphical game, it is equally significant that the control policy is distributed and adheres to Nash equilibrium solution. However, achieving both distributed control and Nash equilibrium simultaneously has proven impossible in most existing game formulations. This paper proposes a new game formulation that can overcome this limitation, enabling each agent to reach a Nash equilibrium under a distributed policy, thereby improving system performance. Furthermore, a partially model-free reinforcement learning method is employed to obtain the optimal policy when the dynamic information is partially unknown, with admissibility condition of the initial policy further relaxed. Finally, two comparative simulations are presented to demonstrate the validity and superiority of the proposed approach.
对抗环境下多智能体博弈的分布式纳什均衡解:一种强化学习方法
本文从微分图形对策的角度研究了具有对抗输入的线性多智能体系统的领导-追随者最优共识问题。对于用微分图对策描述的多智能体最优控制问题,控制策略的分散性和服从纳什均衡解同样重要。然而,在大多数现有的博弈公式中,同时实现分布式控制和纳什均衡已被证明是不可能的。本文提出了一种新的博弈公式,可以克服这一限制,使每个智能体在分布式策略下达到纳什均衡,从而提高系统性能。进一步,采用部分无模型强化学习方法,在动态信息部分未知时获得最优策略,并进一步放宽初始策略的可接受条件。最后,通过两个对比仿真验证了该方法的有效性和优越性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Automatica
Automatica 工程技术-工程:电子与电气
CiteScore
10.70
自引率
7.80%
发文量
617
审稿时长
5 months
期刊介绍: Automatica is a leading archival publication in the field of systems and control. The field encompasses today a broad set of areas and topics, and is thriving not only within itself but also in terms of its impact on other fields, such as communications, computers, biology, energy and economics. Since its inception in 1963, Automatica has kept abreast with the evolution of the field over the years, and has emerged as a leading publication driving the trends in the field. After being founded in 1963, Automatica became a journal of the International Federation of Automatic Control (IFAC) in 1969. It features a characteristic blend of theoretical and applied papers of archival, lasting value, reporting cutting edge research results by authors across the globe. It features articles in distinct categories, including regular, brief and survey papers, technical communiqués, correspondence items, as well as reviews on published books of interest to the readership. It occasionally publishes special issues on emerging new topics or established mature topics of interest to a broad audience. Automatica solicits original high-quality contributions in all the categories listed above, and in all areas of systems and control interpreted in a broad sense and evolving constantly. They may be submitted directly to a subject editor or to the Editor-in-Chief if not sure about the subject area. Editorial procedures in place assure careful, fair, and prompt handling of all submitted articles. Accepted papers appear in the journal in the shortest time feasible given production time constraints.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信