Invited: Efficient Reinforcement Learning for Automating Human Decision-Making in SoC Design

Shankar Sadasivam, Zhuo Chen, Jinwon Lee, Rajeev Jain
{"title":"Invited: Efficient Reinforcement Learning for Automating Human Decision-Making in SoC Design","authors":"Shankar Sadasivam, Zhuo Chen, Jinwon Lee, Rajeev Jain","doi":"10.1145/3195970.3199855","DOIUrl":null,"url":null,"abstract":"The exponential growth in PVT corners due to Moore's law scaling, and the increasing demand for consumer applications and longer battery life in mobile devices, has ushered in significant cost and power-related challenges for designing and productizing mobile chips within a predictable schedule. Two main reasons for this are the reliance on human decision-making to achieve the desired performance within the target area and power budget, and significant increases in complexity of the human decision-making space. The problem is that to-date human design experience has not been replaced by design automation tools, and tasks requiring experience of past designs are still being performed manually.In this paper we investigate how machine learning may be applied to develop tools that learn from experience just like human designers, thus automating tasks that still require human intervention. The potential advantage of the machine learning approach is the ability to scale with increasing complexity and therefore hold the design-time constant with same manpower.Reinforcement Learning (RL) is a machine learning technique that allows us to mimic a human designers' ability to learn from experience and automate human decision-making, without loss in quality of the design, while making the design time independent of the complexity. In this paper we show how manual design tasks can be abstracted as RL problems. Based on the experience with applying RL to one of these problems, we show that RL can automatically achieve results similar to human designs, but in a predictable schedule. However, a major drawback is that the RL solution can require a prohibitively large number of iterations for training. If efficient training techniques can be developed for RL, it holds great promise to automate tasks requiring human experience. In this paper we present a Bayesian Optimization technique for reducing the RL training time.","PeriodicalId":6491,"journal":{"name":"2018 55th ACM/ESDA/IEEE Design Automation Conference (DAC)","volume":"20 1","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 55th ACM/ESDA/IEEE Design Automation Conference (DAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3195970.3199855","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

The exponential growth in PVT corners due to Moore's law scaling, and the increasing demand for consumer applications and longer battery life in mobile devices, has ushered in significant cost and power-related challenges for designing and productizing mobile chips within a predictable schedule. Two main reasons for this are the reliance on human decision-making to achieve the desired performance within the target area and power budget, and significant increases in complexity of the human decision-making space. The problem is that to-date human design experience has not been replaced by design automation tools, and tasks requiring experience of past designs are still being performed manually.In this paper we investigate how machine learning may be applied to develop tools that learn from experience just like human designers, thus automating tasks that still require human intervention. The potential advantage of the machine learning approach is the ability to scale with increasing complexity and therefore hold the design-time constant with same manpower.Reinforcement Learning (RL) is a machine learning technique that allows us to mimic a human designers' ability to learn from experience and automate human decision-making, without loss in quality of the design, while making the design time independent of the complexity. In this paper we show how manual design tasks can be abstracted as RL problems. Based on the experience with applying RL to one of these problems, we show that RL can automatically achieve results similar to human designs, but in a predictable schedule. However, a major drawback is that the RL solution can require a prohibitively large number of iterations for training. If efficient training techniques can be developed for RL, it holds great promise to automate tasks requiring human experience. In this paper we present a Bayesian Optimization technique for reducing the RL training time.
邀请:SoC设计中人类决策自动化的高效强化学习
由于摩尔定律的缩放,PVT角落的指数级增长,以及消费者应用需求的增长和移动设备电池寿命的延长,为在可预测的时间表内设计和生产移动芯片带来了巨大的成本和功耗方面的挑战。造成这种情况的两个主要原因是依靠人工决策来实现目标区域和功率预算内的期望性能,以及人工决策空间的复杂性显着增加。问题是,迄今为止,人类的设计经验并没有被设计自动化工具所取代,需要过去设计经验的任务仍然是手动执行的。在本文中,我们研究了如何将机器学习应用于开发像人类设计师一样从经验中学习的工具,从而使仍然需要人工干预的任务自动化。机器学习方法的潜在优势是能够随着复杂性的增加而扩展,因此在相同的人力条件下保持设计时间不变。强化学习(RL)是一种机器学习技术,它允许我们模仿人类设计师从经验中学习和自动化人类决策的能力,而不会损失设计质量,同时使设计时间与复杂性无关。在本文中,我们展示了如何将手工设计任务抽象为强化学习问题。基于将强化学习应用于其中一个问题的经验,我们表明强化学习可以自动实现类似于人类设计的结果,但在可预测的时间表内。然而,一个主要的缺点是RL解决方案可能需要大量的训练迭代。如果可以为强化学习开发有效的训练技术,那么自动化需要人类经验的任务将大有希望。本文提出了一种减少强化学习训练时间的贝叶斯优化技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信