Finite Approximations for Mean-Field Type Multi-agent Control and Their Near Optimality

IF 1.7 2区 数学 Q2 MATHEMATICS, APPLIED
Erhan Bayraktar, Nicole Bäuerle, Ali Devran Kara
{"title":"Finite Approximations for Mean-Field Type Multi-agent Control and Their Near Optimality","authors":"Erhan Bayraktar,&nbsp;Nicole Bäuerle,&nbsp;Ali Devran Kara","doi":"10.1007/s00245-025-10279-x","DOIUrl":null,"url":null,"abstract":"<div><p>We study a multi-agent mean-field type control problem in discrete time where the agents aim to find a socially optimal strategy and where the state and action spaces for the agents are assumed to be continuous. The agents are only weakly coupled through the distribution of their state variables. The problem in its original form can be formulated as a classical Markov decision process (MDP), however, this formulation suffers from several practical difficulties. In this work, we attempt to overcome the curse of dimensionality, coordination complexity between the agents, and the necessity of perfect feedback collection from all the agents (which might be hard to do for large populations.) We provide several approximations: we establish the near optimality of the action and state space discretization of the agents under standard regularity assumptions for the considered formulation by constructing and studying the measure valued MDP counterpart for finite and infinite population settings. It is a well known approach to consider the infinite population problem for mean-field type models, since it provides symmetric policies for the agents which simplifies the coordination between the agents. However, the optimality analysis is harder as the state space of the measure valued infinite population MDP is continuous (even after space discretization of the agents). Therefore, as a final step, we provide two further approximations for the infinite population problem: the first one directly aggregates the probability measure space, and requires the distribution of the agents to be collected and mapped with a nearest neighbor map, and the second method approximates the measure valued MDP through the empirical distributions of a smaller sized sub-population, for which one only needs keep track of the mean-field term as an estimate by collecting the state information of a small sub-population. For each of the approximation methods, we provide provable regret bounds.</p></div>","PeriodicalId":55566,"journal":{"name":"Applied Mathematics and Optimization","volume":"92 1","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2025-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Mathematics and Optimization","FirstCategoryId":"100","ListUrlMain":"https://link.springer.com/article/10.1007/s00245-025-10279-x","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0

Abstract

We study a multi-agent mean-field type control problem in discrete time where the agents aim to find a socially optimal strategy and where the state and action spaces for the agents are assumed to be continuous. The agents are only weakly coupled through the distribution of their state variables. The problem in its original form can be formulated as a classical Markov decision process (MDP), however, this formulation suffers from several practical difficulties. In this work, we attempt to overcome the curse of dimensionality, coordination complexity between the agents, and the necessity of perfect feedback collection from all the agents (which might be hard to do for large populations.) We provide several approximations: we establish the near optimality of the action and state space discretization of the agents under standard regularity assumptions for the considered formulation by constructing and studying the measure valued MDP counterpart for finite and infinite population settings. It is a well known approach to consider the infinite population problem for mean-field type models, since it provides symmetric policies for the agents which simplifies the coordination between the agents. However, the optimality analysis is harder as the state space of the measure valued infinite population MDP is continuous (even after space discretization of the agents). Therefore, as a final step, we provide two further approximations for the infinite population problem: the first one directly aggregates the probability measure space, and requires the distribution of the agents to be collected and mapped with a nearest neighbor map, and the second method approximates the measure valued MDP through the empirical distributions of a smaller sized sub-population, for which one only needs keep track of the mean-field term as an estimate by collecting the state information of a small sub-population. For each of the approximation methods, we provide provable regret bounds.

平均场型多智能体控制的有限逼近及其近最优性
我们研究了一个离散时间的多智能体平均场型控制问题,其中智能体的目标是寻找社会最优策略,并且假设智能体的状态和动作空间是连续的。代理只是通过状态变量的分布弱耦合的。该问题的原始形式可以表述为经典的马尔可夫决策过程(MDP),然而,这种表述存在一些实际困难。在这项工作中,我们试图克服维度的诅咒,智能体之间的协调复杂性,以及从所有智能体收集完美反馈的必要性(这对于大群体来说可能很难做到)。我们提供了几个近似:我们通过构造和研究有限和无限种群设置下的测度值MDP对偶,在考虑的公式的标准规则假设下,建立了agent的动作和状态空间离散化的近最优性。对于平均域型模型,它是一种众所周知的考虑无限总体问题的方法,因为它为智能体提供了对称策略,简化了智能体之间的协调。然而,由于测度值无限总体MDP的状态空间是连续的(即使在agent的空间离散化之后),因此最优性分析比较困难。因此,作为最后一步,我们为无限人口问题提供了两个进一步的近似:第一种方法直接聚集概率度量空间,需要收集agent的分布并使用最近邻映射;第二种方法通过较小规模子种群的经验分布来逼近度量值MDP,只需通过收集较小子种群的状态信息来跟踪平均场项作为估计。对于每一种近似方法,我们都提供了可证明的遗憾界。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
3.30
自引率
5.60%
发文量
103
审稿时长
>12 weeks
期刊介绍: The Applied Mathematics and Optimization Journal covers a broad range of mathematical methods in particular those that bridge with optimization and have some connection with applications. Core topics include calculus of variations, partial differential equations, stochastic control, optimization of deterministic or stochastic systems in discrete or continuous time, homogenization, control theory, mean field games, dynamic games and optimal transport. Algorithmic, data analytic, machine learning and numerical methods which support the modeling and analysis of optimization problems are encouraged. Of great interest are papers which show some novel idea in either the theory or model which include some connection with potential applications in science and engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信