通过策略嵌入实现代理辅助的进化强化学习

Lan Tang, Xiaxi Li, Jinyuan Zhang, Guiying Li, Peng Yang, Ke Tang
{"title":"通过策略嵌入实现代理辅助的进化强化学习","authors":"Lan Tang, Xiaxi Li, Jinyuan Zhang, Guiying Li, Peng Yang, Ke Tang","doi":"10.48550/arXiv.2301.13374","DOIUrl":null,"url":null,"abstract":"Evolutionary Reinforcement Learning (ERL) that applying Evolutionary Algorithms (EAs) to optimize the weight parameters of Deep Neural Network (DNN) based policies has been widely regarded as an alternative to traditional reinforcement learning methods. However, the evaluation of the iteratively generated population usually requires a large amount of computational time and can be prohibitively expensive, which may potentially restrict the applicability of ERL. Surrogate is often used to reduce the computational burden of evaluation in EAs. Unfortunately, in ERL, each individual of policy usually represents millions of weights parameters of DNN. This high-dimensional representation of policy has introduced a great challenge to the application of surrogates into ERL to speed up training. This paper proposes a PE-SAERL Framework to at the first time enable surrogate-assisted evolutionary reinforcement learning via policy embedding (PE). Empirical results on 5 Atari games show that the proposed method can perform more efficiently than the four state-of-the-art algorithms. The training process is accelerated up to 7x on tested games, comparing to its counterpart without the surrogate and PE.","PeriodicalId":88899,"journal":{"name":"International Conference on Bio-inspired Computing, Theories and Applications : [proceedings]. International Conference on Bio-inspired Computing, Theories and Applications","volume":"41 1","pages":"233-247"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding\",\"authors\":\"Lan Tang, Xiaxi Li, Jinyuan Zhang, Guiying Li, Peng Yang, Ke Tang\",\"doi\":\"10.48550/arXiv.2301.13374\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Evolutionary Reinforcement Learning (ERL) that applying Evolutionary Algorithms (EAs) to optimize the weight parameters of Deep Neural Network (DNN) based policies has been widely regarded as an alternative to traditional reinforcement learning methods. However, the evaluation of the iteratively generated population usually requires a large amount of computational time and can be prohibitively expensive, which may potentially restrict the applicability of ERL. Surrogate is often used to reduce the computational burden of evaluation in EAs. Unfortunately, in ERL, each individual of policy usually represents millions of weights parameters of DNN. This high-dimensional representation of policy has introduced a great challenge to the application of surrogates into ERL to speed up training. This paper proposes a PE-SAERL Framework to at the first time enable surrogate-assisted evolutionary reinforcement learning via policy embedding (PE). Empirical results on 5 Atari games show that the proposed method can perform more efficiently than the four state-of-the-art algorithms. The training process is accelerated up to 7x on tested games, comparing to its counterpart without the surrogate and PE.\",\"PeriodicalId\":88899,\"journal\":{\"name\":\"International Conference on Bio-inspired Computing, Theories and Applications : [proceedings]. International Conference on Bio-inspired Computing, Theories and Applications\",\"volume\":\"41 1\",\"pages\":\"233-247\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Conference on Bio-inspired Computing, Theories and Applications : [proceedings]. International Conference on Bio-inspired Computing, Theories and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.48550/arXiv.2301.13374\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Bio-inspired Computing, Theories and Applications : [proceedings]. International Conference on Bio-inspired Computing, Theories and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2301.13374","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

进化强化学习(ERL)是一种应用进化算法(EAs)来优化基于深度神经网络(DNN)策略的权重参数的方法,被广泛认为是传统强化学习方法的替代方法。然而,对迭代生成的种群进行评估通常需要大量的计算时间,并且可能非常昂贵,这可能会限制ERL的适用性。代理通常用于减少ea评估的计算负担。不幸的是,在ERL中,每个策略个体通常代表数百万个深度神经网络的权重参数。这种对政策的高维表示给代理模型在ERL中的应用带来了很大的挑战。本文首次提出了一个PE- saerl框架,通过策略嵌入(PE)实现代理辅助进化强化学习。5个雅达利游戏的实证结果表明,所提出的方法比四种最先进的算法执行效率更高。在测试游戏中,与没有替代和PE的情况相比,训练过程加快了7倍。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding
Evolutionary Reinforcement Learning (ERL) that applying Evolutionary Algorithms (EAs) to optimize the weight parameters of Deep Neural Network (DNN) based policies has been widely regarded as an alternative to traditional reinforcement learning methods. However, the evaluation of the iteratively generated population usually requires a large amount of computational time and can be prohibitively expensive, which may potentially restrict the applicability of ERL. Surrogate is often used to reduce the computational burden of evaluation in EAs. Unfortunately, in ERL, each individual of policy usually represents millions of weights parameters of DNN. This high-dimensional representation of policy has introduced a great challenge to the application of surrogates into ERL to speed up training. This paper proposes a PE-SAERL Framework to at the first time enable surrogate-assisted evolutionary reinforcement learning via policy embedding (PE). Empirical results on 5 Atari games show that the proposed method can perform more efficiently than the four state-of-the-art algorithms. The training process is accelerated up to 7x on tested games, comparing to its counterpart without the surrogate and PE.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信