{"title":"在GPU上运行强化学习代理进行多人同时博弈仿真","authors":"Koichi Moriyama, Yoshiya Kurogi, Atsuko Mutoh, Tohgoroh Matsui, Nobuhiro Inuzuka","doi":"10.1109/AGENTS.2019.8929206","DOIUrl":null,"url":null,"abstract":"It is desirable for multi-agent simulation to be run in parallel; if many agents run simultaneously, the total run time is reduced. It is popular to use GPGPU technology as an inexpensive parallelizing approach in simulation, but the “agents” runnable on GPU were simple, rule-based ones like elements in a scientific simulation. This work implements more complicated, learning agents on GPU. We consider an environment where many reinforcement learning agents learning their behavior in an iterated two-person simultaneous game while changing peers. It is necessary to run many simulations in each of which a pair of agents play the game. In this work, we implement on GPU the simulations where the agents learn with reinforcement learning and compare two methods assigning the simulations to GPU cores.","PeriodicalId":235878,"journal":{"name":"2019 IEEE International Conference on Agents (ICA)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Running Reinforcement Learning Agents on GPU for Many Simulations of Two-Person Simultaneous Games\",\"authors\":\"Koichi Moriyama, Yoshiya Kurogi, Atsuko Mutoh, Tohgoroh Matsui, Nobuhiro Inuzuka\",\"doi\":\"10.1109/AGENTS.2019.8929206\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"It is desirable for multi-agent simulation to be run in parallel; if many agents run simultaneously, the total run time is reduced. It is popular to use GPGPU technology as an inexpensive parallelizing approach in simulation, but the “agents” runnable on GPU were simple, rule-based ones like elements in a scientific simulation. This work implements more complicated, learning agents on GPU. We consider an environment where many reinforcement learning agents learning their behavior in an iterated two-person simultaneous game while changing peers. It is necessary to run many simulations in each of which a pair of agents play the game. In this work, we implement on GPU the simulations where the agents learn with reinforcement learning and compare two methods assigning the simulations to GPU cores.\",\"PeriodicalId\":235878,\"journal\":{\"name\":\"2019 IEEE International Conference on Agents (ICA)\",\"volume\":\"30 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE International Conference on Agents (ICA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AGENTS.2019.8929206\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE International Conference on Agents (ICA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AGENTS.2019.8929206","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Running Reinforcement Learning Agents on GPU for Many Simulations of Two-Person Simultaneous Games
It is desirable for multi-agent simulation to be run in parallel; if many agents run simultaneously, the total run time is reduced. It is popular to use GPGPU technology as an inexpensive parallelizing approach in simulation, but the “agents” runnable on GPU were simple, rule-based ones like elements in a scientific simulation. This work implements more complicated, learning agents on GPU. We consider an environment where many reinforcement learning agents learning their behavior in an iterated two-person simultaneous game while changing peers. It is necessary to run many simulations in each of which a pair of agents play the game. In this work, we implement on GPU the simulations where the agents learn with reinforcement learning and compare two methods assigning the simulations to GPU cores.