Measuring the Impact of Memory Replay in Training Pacman Agents using Reinforcement Learning

Fabian Fallas-Moya, Jeremiah Duncan, Tabitha K. Samuel, Amir Sadovnik
{"title":"Measuring the Impact of Memory Replay in Training Pacman Agents using Reinforcement Learning","authors":"Fabian Fallas-Moya, Jeremiah Duncan, Tabitha K. Samuel, Amir Sadovnik","doi":"10.1109/CLEI53233.2021.9640031","DOIUrl":null,"url":null,"abstract":"Reinforcement Learning has been widely applied to play classic games where the agents learn the rules by playing the game by themselves. Recent works in general Reinforcement Learning use many improvements such as memory replay to boost the results and training time but we have not found research that focuses on the impact of memory replay in agents that play simple classic video games. In this research, we present an analysis of the impact of three different techniques of memory replay in the performance of a Deep Q-Learning model using different levels of difficulty of the Pacman video game. Also, we propose a multi-channel image - a novel way to create input tensors for training the model - inspired by one-hot encoding, and we show in the experiment section that the performance is improved by using this idea. We find that our model is able to learn faster than previous work and is even able to learn how to consistently win on the mediumClassic board after only 3,000 training episodes, previously thought to take much longer.","PeriodicalId":6803,"journal":{"name":"2021 XLVII Latin American Computing Conference (CLEI)","volume":"11 1","pages":"1-8"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 XLVII Latin American Computing Conference (CLEI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CLEI53233.2021.9640031","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Reinforcement Learning has been widely applied to play classic games where the agents learn the rules by playing the game by themselves. Recent works in general Reinforcement Learning use many improvements such as memory replay to boost the results and training time but we have not found research that focuses on the impact of memory replay in agents that play simple classic video games. In this research, we present an analysis of the impact of three different techniques of memory replay in the performance of a Deep Q-Learning model using different levels of difficulty of the Pacman video game. Also, we propose a multi-channel image - a novel way to create input tensors for training the model - inspired by one-hot encoding, and we show in the experiment section that the performance is improved by using this idea. We find that our model is able to learn faster than previous work and is even able to learn how to consistently win on the mediumClassic board after only 3,000 training episodes, previously thought to take much longer.
用强化学习测量记忆重放在训练吃豆人代理中的影响
强化学习已经被广泛应用于经典游戏中,在经典游戏中,智能体通过自己玩游戏来学习规则。最近在一般强化学习方面的工作使用了许多改进,例如记忆重播来提高结果和训练时间,但我们还没有发现专注于记忆重播对玩简单经典电子游戏的智能体的影响的研究。在这项研究中,我们分析了三种不同的记忆回放技术对深度q -学习模型性能的影响,这些模型使用了不同难度的吃豆人视频游戏。此外,我们还提出了一种多通道图像——一种受单热编码启发创建用于训练模型的输入张量的新方法,并且我们在实验部分中表明,使用这种思想可以提高性能。我们发现我们的模型能够比以前的工作更快地学习,甚至能够在只经过3000次训练后就学会如何在mediumClassic棋盘上持续获胜,而以前认为这需要更长的时间。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信