{"title":"一个蜂群每个女王:随机游戏的粒子群学习","authors":"Alain Tcheukam Siwe, H. Tembine","doi":"10.1109/SASO.2016.22","DOIUrl":null,"url":null,"abstract":"This article examines a particle swarm collaborative model-free learning algorithm for approximating equilibria of stochastic games with continuous action spaces. The results support the argument that a simple learning algorithm which consists to explore the continuous action set by means of multi-population of particles can provide a satisfactory solution. A collaborative learning between the particles of the same player takes place during the interactions of the game, in which the players and the particles have no direct knowledge of the payoff model. Each particle is allowed to observe her own payoff and has only one-step memory. The existing results linking the outcomes to stationary satisfactory set do not apply to this situation because of continuous action space and non-convex local response. We provide a different approach to stochastic differential inclusion for arbitrary number of agents.","PeriodicalId":383753,"journal":{"name":"2016 IEEE 10th International Conference on Self-Adaptive and Self-Organizing Systems (SASO)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"One Swarm per Queen: A Particle Swarm Learning for Stochastic Games\",\"authors\":\"Alain Tcheukam Siwe, H. Tembine\",\"doi\":\"10.1109/SASO.2016.22\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This article examines a particle swarm collaborative model-free learning algorithm for approximating equilibria of stochastic games with continuous action spaces. The results support the argument that a simple learning algorithm which consists to explore the continuous action set by means of multi-population of particles can provide a satisfactory solution. A collaborative learning between the particles of the same player takes place during the interactions of the game, in which the players and the particles have no direct knowledge of the payoff model. Each particle is allowed to observe her own payoff and has only one-step memory. The existing results linking the outcomes to stationary satisfactory set do not apply to this situation because of continuous action space and non-convex local response. We provide a different approach to stochastic differential inclusion for arbitrary number of agents.\",\"PeriodicalId\":383753,\"journal\":{\"name\":\"2016 IEEE 10th International Conference on Self-Adaptive and Self-Organizing Systems (SASO)\",\"volume\":\"69 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE 10th International Conference on Self-Adaptive and Self-Organizing Systems (SASO)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SASO.2016.22\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE 10th International Conference on Self-Adaptive and Self-Organizing Systems (SASO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SASO.2016.22","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
One Swarm per Queen: A Particle Swarm Learning for Stochastic Games
This article examines a particle swarm collaborative model-free learning algorithm for approximating equilibria of stochastic games with continuous action spaces. The results support the argument that a simple learning algorithm which consists to explore the continuous action set by means of multi-population of particles can provide a satisfactory solution. A collaborative learning between the particles of the same player takes place during the interactions of the game, in which the players and the particles have no direct knowledge of the payoff model. Each particle is allowed to observe her own payoff and has only one-step memory. The existing results linking the outcomes to stationary satisfactory set do not apply to this situation because of continuous action space and non-convex local response. We provide a different approach to stochastic differential inclusion for arbitrary number of agents.