{"title":"MCTS与影响地图一般视频游戏玩","authors":"Hyun-Soo Park, Kyung-Joong Kim","doi":"10.1109/CIG.2015.7317896","DOIUrl":null,"url":null,"abstract":"In the General Video Game-AI competition in 2014 IEEE Computational Intelligence in Games, Monte Carlo Tree Search (MCTS) outperformed other alternatives. Interestingly, the sample MCTS ranked in the third place. However, MCTS was not always perfect in this problem. For example, it cannot explore enough search space of video games because of time constraints. As a result, if the AI player receives only limited rewards from game environments, it is likely to lose the way and moves almost randomly. In this paper, we propose to use influence map (IM), a numerical representation of influence on the game map, to find a road to rewards over the horizon. We reported average winning ratio improvement over alternatives and successful/unsuccessful cases of our algorithm.","PeriodicalId":244862,"journal":{"name":"2015 IEEE Conference on Computational Intelligence and Games (CIG)","volume":"136 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"16","resultStr":"{\"title\":\"MCTS with influence map for general video game playing\",\"authors\":\"Hyun-Soo Park, Kyung-Joong Kim\",\"doi\":\"10.1109/CIG.2015.7317896\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the General Video Game-AI competition in 2014 IEEE Computational Intelligence in Games, Monte Carlo Tree Search (MCTS) outperformed other alternatives. Interestingly, the sample MCTS ranked in the third place. However, MCTS was not always perfect in this problem. For example, it cannot explore enough search space of video games because of time constraints. As a result, if the AI player receives only limited rewards from game environments, it is likely to lose the way and moves almost randomly. In this paper, we propose to use influence map (IM), a numerical representation of influence on the game map, to find a road to rewards over the horizon. We reported average winning ratio improvement over alternatives and successful/unsuccessful cases of our algorithm.\",\"PeriodicalId\":244862,\"journal\":{\"name\":\"2015 IEEE Conference on Computational Intelligence and Games (CIG)\",\"volume\":\"136 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-11-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"16\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2015 IEEE Conference on Computational Intelligence and Games (CIG)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CIG.2015.7317896\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE Conference on Computational Intelligence and Games (CIG)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIG.2015.7317896","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
MCTS with influence map for general video game playing
In the General Video Game-AI competition in 2014 IEEE Computational Intelligence in Games, Monte Carlo Tree Search (MCTS) outperformed other alternatives. Interestingly, the sample MCTS ranked in the third place. However, MCTS was not always perfect in this problem. For example, it cannot explore enough search space of video games because of time constraints. As a result, if the AI player receives only limited rewards from game environments, it is likely to lose the way and moves almost randomly. In this paper, we propose to use influence map (IM), a numerical representation of influence on the game map, to find a road to rewards over the horizon. We reported average winning ratio improvement over alternatives and successful/unsuccessful cases of our algorithm.