严格纳什均衡博弈中有承诺参与者的随机自适应学习

Naoki Funai
{"title":"严格纳什均衡博弈中有承诺参与者的随机自适应学习","authors":"Naoki Funai","doi":"10.2139/ssrn.3944342","DOIUrl":null,"url":null,"abstract":"In this paper, we investigate the condition under which players of an adaptive learning model, including the one of stochastic fictitious play learning, learn to follow a logit quantal response equilibrium corresponding to a strict Nash equilibrium, the one which approaches the Nash equilibrium as the noise term approaches zero. In particular, we consider the situation in which adaptive players face a fixed normal form game repeatedly and may interact with committed players, who do not revise their behaviour and follow a specific strategy in each period. When committed players follow a logit quantal response equilibrium corresponding to a strict Nash equilibrium and the probability of each adaptive player interacting with committed players is large enough, we show that adaptive players learn to follow the equilibrium that committed players follow almost surely; we also show that any strict Nash equilibrium can be logit quantal response equilibrium approachable so that we can pick any strict Nash equilibrium for committed players to follow. Also, we show that when the probability of interacting with committed players is small, players of a more general adaptive learning model, including the ones of stochastic fictitious play learning, payoff assessment learning and delta learning may learn to follow an equilibrium different from the one committed players follow with positive probability. Lastly, we also consider the case in which there do not exist committed players and show that when players of the general adaptive learning model have enough experience and their behaviour is close enough to a logit quantal response equilibrium corresponding to a strict Nash equilibrium, then their behaviour converges to the equilibrium with probability close to one.","PeriodicalId":259640,"journal":{"name":"Information Theory & Research eJournal","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Stochastic Adaptive Learning With Committed Players in Games With Strict Nash Equilibria\",\"authors\":\"Naoki Funai\",\"doi\":\"10.2139/ssrn.3944342\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we investigate the condition under which players of an adaptive learning model, including the one of stochastic fictitious play learning, learn to follow a logit quantal response equilibrium corresponding to a strict Nash equilibrium, the one which approaches the Nash equilibrium as the noise term approaches zero. In particular, we consider the situation in which adaptive players face a fixed normal form game repeatedly and may interact with committed players, who do not revise their behaviour and follow a specific strategy in each period. When committed players follow a logit quantal response equilibrium corresponding to a strict Nash equilibrium and the probability of each adaptive player interacting with committed players is large enough, we show that adaptive players learn to follow the equilibrium that committed players follow almost surely; we also show that any strict Nash equilibrium can be logit quantal response equilibrium approachable so that we can pick any strict Nash equilibrium for committed players to follow. Also, we show that when the probability of interacting with committed players is small, players of a more general adaptive learning model, including the ones of stochastic fictitious play learning, payoff assessment learning and delta learning may learn to follow an equilibrium different from the one committed players follow with positive probability. Lastly, we also consider the case in which there do not exist committed players and show that when players of the general adaptive learning model have enough experience and their behaviour is close enough to a logit quantal response equilibrium corresponding to a strict Nash equilibrium, then their behaviour converges to the equilibrium with probability close to one.\",\"PeriodicalId\":259640,\"journal\":{\"name\":\"Information Theory & Research eJournal\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Theory & Research eJournal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2139/ssrn.3944342\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Theory & Research eJournal","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2139/ssrn.3944342","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

本文研究了自适应学习模型(包括随机虚拟游戏学习模型)的参与者学习遵循严格纳什均衡对应的logit量子响应均衡的条件,该均衡在噪声项趋近于零时趋近于纳什均衡。特别是,我们考虑的情况是,适应性玩家反复面对固定的正常形式的游戏,并可能与忠诚的玩家互动,他们不会改变自己的行为,并在每个时期遵循特定的策略。当承诺参与者遵循对应于严格纳什均衡的logit量子响应均衡,且每个自适应参与者与承诺参与者互动的概率足够大时,我们发现自适应参与者学习遵循承诺参与者几乎肯定遵循的均衡;我们还证明了任何严格纳什均衡都可以是可接近的logit量子响应均衡,这样我们就可以为承诺的参与者选择任何严格纳什均衡。此外,我们还表明,当与承诺参与者互动的概率很小时,更一般的适应性学习模型(包括随机虚拟游戏学习、收益评估学习和delta学习)的参与者可能会学习遵循与承诺参与者遵循的正概率均衡不同的均衡。最后,我们还考虑了不存在承诺参与者的情况,并表明当一般自适应学习模型的参与者有足够的经验并且他们的行为足够接近严格纳什均衡对应的logit量子响应均衡时,他们的行为收敛于概率接近1的均衡。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Stochastic Adaptive Learning With Committed Players in Games With Strict Nash Equilibria
In this paper, we investigate the condition under which players of an adaptive learning model, including the one of stochastic fictitious play learning, learn to follow a logit quantal response equilibrium corresponding to a strict Nash equilibrium, the one which approaches the Nash equilibrium as the noise term approaches zero. In particular, we consider the situation in which adaptive players face a fixed normal form game repeatedly and may interact with committed players, who do not revise their behaviour and follow a specific strategy in each period. When committed players follow a logit quantal response equilibrium corresponding to a strict Nash equilibrium and the probability of each adaptive player interacting with committed players is large enough, we show that adaptive players learn to follow the equilibrium that committed players follow almost surely; we also show that any strict Nash equilibrium can be logit quantal response equilibrium approachable so that we can pick any strict Nash equilibrium for committed players to follow. Also, we show that when the probability of interacting with committed players is small, players of a more general adaptive learning model, including the ones of stochastic fictitious play learning, payoff assessment learning and delta learning may learn to follow an equilibrium different from the one committed players follow with positive probability. Lastly, we also consider the case in which there do not exist committed players and show that when players of the general adaptive learning model have enough experience and their behaviour is close enough to a logit quantal response equilibrium corresponding to a strict Nash equilibrium, then their behaviour converges to the equilibrium with probability close to one.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信