{"title":"Umbrella Reinforcement Learning – computationally efficient tool for hard non-linear problems","authors":"Egor E. Nuzhin , Nikolay V. Brilliantov","doi":"10.1016/j.cnsns.2024.108583","DOIUrl":null,"url":null,"abstract":"<div><div>We report a novel, computationally efficient approach for solving hard nonlinear problems of reinforcement learning (RL). Here we combine umbrella sampling, from computational physics/chemistry, with optimal control methods. The approach is realized on the basis of neural networks, with the use of policy gradient. It outperforms, by computational efficiency and implementation universality, all available state-of-the-art algorithms, in application to hard RL problems with sparse reward, state traps and lack of terminal states. The proposed approach uses an ensemble of simultaneously acting agents, with a modified reward which includes the ensemble entropy, yielding an optimal exploration-exploitation balance.</div></div>","PeriodicalId":50658,"journal":{"name":"Communications in Nonlinear Science and Numerical Simulation","volume":"143 ","pages":"Article 108583"},"PeriodicalIF":3.4000,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Communications in Nonlinear Science and Numerical Simulation","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1007570424007688","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0
Abstract
We report a novel, computationally efficient approach for solving hard nonlinear problems of reinforcement learning (RL). Here we combine umbrella sampling, from computational physics/chemistry, with optimal control methods. The approach is realized on the basis of neural networks, with the use of policy gradient. It outperforms, by computational efficiency and implementation universality, all available state-of-the-art algorithms, in application to hard RL problems with sparse reward, state traps and lack of terminal states. The proposed approach uses an ensemble of simultaneously acting agents, with a modified reward which includes the ensemble entropy, yielding an optimal exploration-exploitation balance.
期刊介绍:
The journal publishes original research findings on experimental observation, mathematical modeling, theoretical analysis and numerical simulation, for more accurate description, better prediction or novel application, of nonlinear phenomena in science and engineering. It offers a venue for researchers to make rapid exchange of ideas and techniques in nonlinear science and complexity.
The submission of manuscripts with cross-disciplinary approaches in nonlinear science and complexity is particularly encouraged.
Topics of interest:
Nonlinear differential or delay equations, Lie group analysis and asymptotic methods, Discontinuous systems, Fractals, Fractional calculus and dynamics, Nonlinear effects in quantum mechanics, Nonlinear stochastic processes, Experimental nonlinear science, Time-series and signal analysis, Computational methods and simulations in nonlinear science and engineering, Control of dynamical systems, Synchronization, Lyapunov analysis, High-dimensional chaos and turbulence, Chaos in Hamiltonian systems, Integrable systems and solitons, Collective behavior in many-body systems, Biological physics and networks, Nonlinear mechanical systems, Complex systems and complexity.
No length limitation for contributions is set, but only concisely written manuscripts are published. Brief papers are published on the basis of Rapid Communications. Discussions of previously published papers are welcome.