Deep Q-Learning for Nash Equilibria: Nash-DQN

Q3 Mathematics
P. Casgrain, Brian Ning, S. Jaimungal
{"title":"Deep Q-Learning for Nash Equilibria: Nash-DQN","authors":"P. Casgrain, Brian Ning, S. Jaimungal","doi":"10.1080/1350486X.2022.2136727","DOIUrl":null,"url":null,"abstract":"ABSTRACT Model-free learning for multi-agent stochastic games is an active area of research. Existing reinforcement learning algorithms, however, are often restricted to zero-sum games and are applicable only in small state-action spaces or other simplified settings. Here, we develop a new data-efficient Deep-Q-learning methodology for model-free learning of Nash equilibria for general-sum stochastic games. The algorithm uses a locally linear-quadratic expansion of the stochastic game, which leads to analytically solvable optimal actions. The expansion is parametrized by deep neural networks to give it sufficient flexibility to learn the environment without the need to experience all state-action pairs. We study symmetry properties of the algorithm stemming from label-invariant stochastic games and as a proof of concept, apply our algorithm to learning optimal trading strategies in competitive electronic markets.","PeriodicalId":35818,"journal":{"name":"Applied Mathematical Finance","volume":"85 1","pages":"62 - 78"},"PeriodicalIF":0.0000,"publicationDate":"2019-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"19","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Mathematical Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/1350486X.2022.2136727","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Mathematics","Score":null,"Total":0}
引用次数: 19

Abstract

ABSTRACT Model-free learning for multi-agent stochastic games is an active area of research. Existing reinforcement learning algorithms, however, are often restricted to zero-sum games and are applicable only in small state-action spaces or other simplified settings. Here, we develop a new data-efficient Deep-Q-learning methodology for model-free learning of Nash equilibria for general-sum stochastic games. The algorithm uses a locally linear-quadratic expansion of the stochastic game, which leads to analytically solvable optimal actions. The expansion is parametrized by deep neural networks to give it sufficient flexibility to learn the environment without the need to experience all state-action pairs. We study symmetry properties of the algorithm stemming from label-invariant stochastic games and as a proof of concept, apply our algorithm to learning optimal trading strategies in competitive electronic markets.
纳什均衡的深度q学习:Nash- dqn
多智能体随机博弈的无模型学习是一个活跃的研究领域。然而,现有的强化学习算法通常仅限于零和游戏,并且仅适用于小的状态-行动空间或其他简化设置。在这里,我们开发了一种新的数据高效的深度q学习方法,用于一般和随机博弈的纳什均衡的无模型学习。该算法对随机对策进行局部线性二次展开,得到解析可解的最优对策。扩展由深度神经网络参数化,使其具有足够的灵活性来学习环境,而无需经历所有状态-动作对。我们研究了源自标签不变随机博弈的算法的对称性,并作为概念证明,将我们的算法应用于竞争性电子市场中学习最优交易策略。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Applied Mathematical Finance
Applied Mathematical Finance Economics, Econometrics and Finance-Finance
CiteScore
2.30
自引率
0.00%
发文量
6
期刊介绍: The journal encourages the confident use of applied mathematics and mathematical modelling in finance. The journal publishes papers on the following: •modelling of financial and economic primitives (interest rates, asset prices etc); •modelling market behaviour; •modelling market imperfections; •pricing of financial derivative securities; •hedging strategies; •numerical methods; •financial engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信