Sibylle D. Müller, N. Schraudolph, P. Koumoutsakos
{"title":"Step size adaptation in evolution strategies using reinforcement learning","authors":"Sibylle D. Müller, N. Schraudolph, P. Koumoutsakos","doi":"10.1109/CEC.2002.1006225","DOIUrl":null,"url":null,"abstract":"We discuss the implementation of a learning algorithm for determining adaptation parameters in evolution strategies. As an initial test case, we consider the application of reinforcement learning for determining the relationship between success rates and the adaptation of step sizes in the (1+1)-evolution strategy. The results from the new adaptive scheme when applied to several test functions are compared with those obtained from the (1+1)-evolution strategy with a priori selected parameters. Our results indicate that assigning good reward measures seems to be crucial to the performance of the combined strategy.","PeriodicalId":184547,"journal":{"name":"Proceedings of the 2002 Congress on Evolutionary Computation. CEC'02 (Cat. No.02TH8600)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2002-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"43","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2002 Congress on Evolutionary Computation. CEC'02 (Cat. No.02TH8600)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CEC.2002.1006225","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 43
Abstract
We discuss the implementation of a learning algorithm for determining adaptation parameters in evolution strategies. As an initial test case, we consider the application of reinforcement learning for determining the relationship between success rates and the adaptation of step sizes in the (1+1)-evolution strategy. The results from the new adaptive scheme when applied to several test functions are compared with those obtained from the (1+1)-evolution strategy with a priori selected parameters. Our results indicate that assigning good reward measures seems to be crucial to the performance of the combined strategy.