{"title":"Parametric Circuit Optimization with Reinforcement Learning","authors":"Changcheng Tang, Zuochang Ye, Yan Wang","doi":"10.1109/ISVLSI.2018.00045","DOIUrl":null,"url":null,"abstract":"In this paper, we focus on solving parametric optimization problems. Such kind of problems is very commonly seen in reality. We propose an efficient method to train a model that connects the solution to the parameters and thus solve all the problems with the same structure and different parameters at the same time. During the training process, instead of solving a series of optimization problems with randomly sampled w independently, we adopt reinforcement learning to accelerate the training process. Two networks are trained alternately. The first network is a value network, and it is trained to fit the target loss function. The second network is a policy network, whose output is connected to the input of the value network and it is trained to minimize the output of the value network. Experiments demonstrate the effectiveness of the proposed method.","PeriodicalId":114330,"journal":{"name":"2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISVLSI.2018.00045","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
In this paper, we focus on solving parametric optimization problems. Such kind of problems is very commonly seen in reality. We propose an efficient method to train a model that connects the solution to the parameters and thus solve all the problems with the same structure and different parameters at the same time. During the training process, instead of solving a series of optimization problems with randomly sampled w independently, we adopt reinforcement learning to accelerate the training process. Two networks are trained alternately. The first network is a value network, and it is trained to fit the target loss function. The second network is a policy network, whose output is connected to the input of the value network and it is trained to minimize the output of the value network. Experiments demonstrate the effectiveness of the proposed method.