Gustavo C. Mangold, Mendeli H. Vainstein, Heitor C.M. Fernandes
{"title":"基于强化学习的空间囚徒困境中的稀释、扩散与共生","authors":"Gustavo C. Mangold, Mendeli H. Vainstein, Heitor C.M. Fernandes","doi":"10.1016/j.chaos.2025.117382","DOIUrl":null,"url":null,"abstract":"<div><div>Recent studies on spatial prisoner’s dilemma games with reinforcement learning have shown that static agents can learn to cooperate through a variety of mechanisms, including noise injection, different learning algorithms, and access to neighbours’ payoff information. In this work, we use an independent multi-agent Q-learning algorithm to investigate the effects of dilution and mobility in the spatial version of the prisoner’s dilemma. Within this framework, different possible actions for the algorithm are defined, linking our results to those of the classical, non-reinforcement learning spatial prisoner’s dilemma. This highlights the algorithm’s versatility in modelling diverse game-theoretical scenarios and demonstrates its potential as a benchmarking tool. Our findings reveal a range of effects, including evidence that games with fixed update rules can be qualitatively equivalent to those with learned ones. Additionally, we observe the emergence of a symbiotic mutualistic effect between populations when multiple actions are defined.</div></div>","PeriodicalId":9764,"journal":{"name":"Chaos Solitons & Fractals","volume":"201 ","pages":"Article 117382"},"PeriodicalIF":5.6000,"publicationDate":"2025-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Dilution, diffusion and symbiosis in spatial prisoner’s dilemma with reinforcement learning\",\"authors\":\"Gustavo C. Mangold, Mendeli H. Vainstein, Heitor C.M. Fernandes\",\"doi\":\"10.1016/j.chaos.2025.117382\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Recent studies on spatial prisoner’s dilemma games with reinforcement learning have shown that static agents can learn to cooperate through a variety of mechanisms, including noise injection, different learning algorithms, and access to neighbours’ payoff information. In this work, we use an independent multi-agent Q-learning algorithm to investigate the effects of dilution and mobility in the spatial version of the prisoner’s dilemma. Within this framework, different possible actions for the algorithm are defined, linking our results to those of the classical, non-reinforcement learning spatial prisoner’s dilemma. This highlights the algorithm’s versatility in modelling diverse game-theoretical scenarios and demonstrates its potential as a benchmarking tool. Our findings reveal a range of effects, including evidence that games with fixed update rules can be qualitatively equivalent to those with learned ones. Additionally, we observe the emergence of a symbiotic mutualistic effect between populations when multiple actions are defined.</div></div>\",\"PeriodicalId\":9764,\"journal\":{\"name\":\"Chaos Solitons & Fractals\",\"volume\":\"201 \",\"pages\":\"Article 117382\"},\"PeriodicalIF\":5.6000,\"publicationDate\":\"2025-10-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Chaos Solitons & Fractals\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0960077925013955\",\"RegionNum\":1,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Chaos Solitons & Fractals","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0960077925013955","RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
Dilution, diffusion and symbiosis in spatial prisoner’s dilemma with reinforcement learning
Recent studies on spatial prisoner’s dilemma games with reinforcement learning have shown that static agents can learn to cooperate through a variety of mechanisms, including noise injection, different learning algorithms, and access to neighbours’ payoff information. In this work, we use an independent multi-agent Q-learning algorithm to investigate the effects of dilution and mobility in the spatial version of the prisoner’s dilemma. Within this framework, different possible actions for the algorithm are defined, linking our results to those of the classical, non-reinforcement learning spatial prisoner’s dilemma. This highlights the algorithm’s versatility in modelling diverse game-theoretical scenarios and demonstrates its potential as a benchmarking tool. Our findings reveal a range of effects, including evidence that games with fixed update rules can be qualitatively equivalent to those with learned ones. Additionally, we observe the emergence of a symbiotic mutualistic effect between populations when multiple actions are defined.
期刊介绍:
Chaos, Solitons & Fractals strives to establish itself as a premier journal in the interdisciplinary realm of Nonlinear Science, Non-equilibrium, and Complex Phenomena. It welcomes submissions covering a broad spectrum of topics within this field, including dynamics, non-equilibrium processes in physics, chemistry, and geophysics, complex matter and networks, mathematical models, computational biology, applications to quantum and mesoscopic phenomena, fluctuations and random processes, self-organization, and social phenomena.