Hyunwook Park, Minsu Kim, Subin Kim, Seungtaek Jeong, Seongguk Kim, Hyungmin Kang, Keunwoo Kim, Keeyoung Son, Gapyeol Park, Kyungjune Son, Taein Shin, Joungho Kim
{"title":"Policy Gradient Reinforcement Learning-based Optimal Decoupling Capacitor Design Method for 2.5-D/3-D ICs using Transformer Network","authors":"Hyunwook Park, Minsu Kim, Subin Kim, Seungtaek Jeong, Seongguk Kim, Hyungmin Kang, Keunwoo Kim, Keeyoung Son, Gapyeol Park, Kyungjune Son, Taein Shin, Joungho Kim","doi":"10.1109/EDAPS50281.2020.9312908","DOIUrl":null,"url":null,"abstract":"In this paper, we first propose a policy gradient reinforcement learning (RL)-based optimal decoupling capacitor (decap) design method for 2.5-D/3-D integrated circuits (ICs) using a transformer network. The proposed method can provide an optimal decap design that meets target impedance. Unlike previous value-based RL methods with simple value approximators such as multi-layer perceptron (MLP) and convolutional neural network (CNN), the proposed method directly parameterizes policy using an attention-based transformer network model. The model is trained through the policy gradient algorithm so that it can achieve larger action space, i.e. search space. For verification, we applied the proposed method to a test hierarchical power distribution network (PDN). We compared convergence results depending on the action space with the previous value-based RL method. As a result, it is validated that the proposed method can cover ×4 times larger action space than that of the previous work.","PeriodicalId":137699,"journal":{"name":"2020 IEEE Electrical Design of Advanced Packaging and Systems (EDAPS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Electrical Design of Advanced Packaging and Systems (EDAPS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EDAPS50281.2020.9312908","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
In this paper, we first propose a policy gradient reinforcement learning (RL)-based optimal decoupling capacitor (decap) design method for 2.5-D/3-D integrated circuits (ICs) using a transformer network. The proposed method can provide an optimal decap design that meets target impedance. Unlike previous value-based RL methods with simple value approximators such as multi-layer perceptron (MLP) and convolutional neural network (CNN), the proposed method directly parameterizes policy using an attention-based transformer network model. The model is trained through the policy gradient algorithm so that it can achieve larger action space, i.e. search space. For verification, we applied the proposed method to a test hierarchical power distribution network (PDN). We compared convergence results depending on the action space with the previous value-based RL method. As a result, it is validated that the proposed method can cover ×4 times larger action space than that of the previous work.
在本文中,我们首先提出了一种基于策略梯度强化学习(RL)的2.5 d /3-D集成电路(ic)的最佳去耦电容(decap)设计方法,该方法使用变压器网络。该方法可以提供满足目标阻抗的最优封装设计。与以往基于值的RL方法使用简单的值逼近器(如多层感知器(MLP)和卷积神经网络(CNN))不同,该方法使用基于注意力的变压器网络模型直接参数化策略。通过策略梯度算法对模型进行训练,使其达到更大的动作空间,即搜索空间。为了验证,我们将该方法应用于一个测试的分层配电网络(PDN)。我们将基于动作空间的收敛结果与之前基于值的RL方法进行了比较。结果验证了所提方法覆盖的动作空间比以往的工作大×4倍。