Nash Q-Learning Agents in Hotelling’s Model: Reestablishing Equilibrium

Jan Vainer, J. Kukacka
{"title":"Nash Q-Learning Agents in Hotelling’s Model: Reestablishing Equilibrium","authors":"Jan Vainer, J. Kukacka","doi":"10.2139/ssrn.3298510","DOIUrl":null,"url":null,"abstract":"This paper examines adaptive agents' behavior in a stochastic dynamic version of the Hotelling's location model. We conduct an agent-based numerical simulation under the Hotelling's setting with two agents who use the Nash Q-learning mechanism for adaptation. This allows exploring what alternations this technique brings compared to the original analytic solution of the famous static game-theoretic model with strong assumptions imposed on players. We discover that under the Nash Q-learning and quadratic consumer cost function, agents with high enough valuation of future profits learn behavior similar to aggressive market strategy. Both agents make similar products and lead a price war to eliminate their opponent from the market. This behavior closely resembles the Principle of Minimum Differentiation from Hotelling's original paper with linear consumer costs. However, the quadratic consumer cost function would otherwise result in the maximum differentiation of production in the original model. Thus, the Principle of Minimum Differentiation can be justified based on repeated interactions of the agents and long-run optimization.","PeriodicalId":431230,"journal":{"name":"ERN: Consumption","volume":"53 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ERN: Consumption","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2139/ssrn.3298510","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

This paper examines adaptive agents' behavior in a stochastic dynamic version of the Hotelling's location model. We conduct an agent-based numerical simulation under the Hotelling's setting with two agents who use the Nash Q-learning mechanism for adaptation. This allows exploring what alternations this technique brings compared to the original analytic solution of the famous static game-theoretic model with strong assumptions imposed on players. We discover that under the Nash Q-learning and quadratic consumer cost function, agents with high enough valuation of future profits learn behavior similar to aggressive market strategy. Both agents make similar products and lead a price war to eliminate their opponent from the market. This behavior closely resembles the Principle of Minimum Differentiation from Hotelling's original paper with linear consumer costs. However, the quadratic consumer cost function would otherwise result in the maximum differentiation of production in the original model. Thus, the Principle of Minimum Differentiation can be justified based on repeated interactions of the agents and long-run optimization.
Hotelling模型中的Nash Q-Learning agent:重新建立均衡
本文在霍特林区位模型的随机动态版本中考察了自适应主体的行为。我们在Hotelling模型下对两个使用Nash q -学习机制进行适应的agent进行了基于agent的数值模拟。这让我们能够探索这种技术与著名的静态博弈论模型的原始分析解决方案(游戏邦注:该模型对玩家施加了强大的假设)相比所带来的改变。我们发现,在纳什q -学习和二次消费者成本函数下,对未来利润估值足够高的代理学习的行为类似于激进市场策略。两家代理商都生产类似的产品,并发动价格战,将对手从市场上排挤出去。这种行为非常类似于Hotelling原始论文中消费者成本为线性的最小微分原理。然而,在原模型中,二次消费者成本函数会导致最大的生产差异。因此,基于代理的重复交互和长期优化,可以证明最小微分原则是合理的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信