{"title":"ML-based Reinforcement Learning Approach for Power Management in SoCs","authors":"D. Akselrod","doi":"10.1109/SOCC46988.2019.1570548498","DOIUrl":null,"url":null,"abstract":"This paper presents a machine learning-based reinforcement learning approach, mapping Finite State Machines, traditionally used for power management control in SoCs, to Markov Decision Process (MDP)-based agents for controlling power management features of Integrated Circuits with application to complex multiprocessor-based SoCs such as CPUs, APUs and GPUs. We present the problem of decision-based control of a number of power management features in ICs consisting of numerous heterogeneous IPs. An infinite-horizon fully observable MDPs are utilized to obtain a policy of actions maximizing the expectation of the formulated Power Management utility function. The approach balances the demand for desired performance while providing an optimal power saving as opposed to commonly used FSM-based power management techniques. MDP framework was employed for power management decision-making under conditions of uncertainly for reinforcement learning. We describe in detail converting power management FSMs into infinite-horizon fully observable MDPs. The approach optimizes itself using reinforcement learning based on specified reward structure and previous performance, yielding an optimal and dynamically adjusted power management mechanism in respect to the formulated model.","PeriodicalId":253998,"journal":{"name":"2019 32nd IEEE International System-on-Chip Conference (SOCC)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 32nd IEEE International System-on-Chip Conference (SOCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SOCC46988.2019.1570548498","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper presents a machine learning-based reinforcement learning approach, mapping Finite State Machines, traditionally used for power management control in SoCs, to Markov Decision Process (MDP)-based agents for controlling power management features of Integrated Circuits with application to complex multiprocessor-based SoCs such as CPUs, APUs and GPUs. We present the problem of decision-based control of a number of power management features in ICs consisting of numerous heterogeneous IPs. An infinite-horizon fully observable MDPs are utilized to obtain a policy of actions maximizing the expectation of the formulated Power Management utility function. The approach balances the demand for desired performance while providing an optimal power saving as opposed to commonly used FSM-based power management techniques. MDP framework was employed for power management decision-making under conditions of uncertainly for reinforcement learning. We describe in detail converting power management FSMs into infinite-horizon fully observable MDPs. The approach optimizes itself using reinforcement learning based on specified reward structure and previous performance, yielding an optimal and dynamically adjusted power management mechanism in respect to the formulated model.