Mayo: A Framework for Auto-generating Hardware Friendly Deep Neural Networks

Yiren Zhao, Xitong Gao, R. Mullins, Chengzhong Xu
{"title":"Mayo: A Framework for Auto-generating Hardware Friendly Deep Neural Networks","authors":"Yiren Zhao, Xitong Gao, R. Mullins, Chengzhong Xu","doi":"10.1145/3212725.3212726","DOIUrl":null,"url":null,"abstract":"Deep Neural Networks (DNNs) have proved to be a convenient and powerful tool for a wide range of problems. However, the extensive computational and memory resource requirements hinder the adoption of DNNs in resource-constrained scenarios. Existing compression methods have been shown to significantly reduce the computation and memory requirements of many popular DNNs. These methods, however, remain elusive to non-experts, as they demand extensive manual tuning of hyperparameters. The effects of combining various compression techniques lack exploration because of the large design space. To alleviate these challenges, this paper proposes an automated framework, Mayo, which is built on top of TensorFlow and can compress DNNs with minimal human intervention. First, we present overriders which are recursively-compositional and can be configured to effectively compress individual components (e.g. weights, biases, layer computations and gradients) in a DNN. Second, we introduce novel heuristics and a global search algorithm to efficiently optimize hyperparameters. We demonstrate that without any manual tuning, Mayo generates a sparse ResNet-18 that is 5.13x smaller than the baseline with no loss in test accuracy. By composing multiple overriders, our tool produces a sparse 6-bit CIFAR-10 classifier with only 0.16% top-1 accuracy loss and a 34x compression rate. Mayo and all compressed models are publicly available. To our knowledge, Mayo is the first framework that supports overlapping multiple compression techniques and automatically optimizes hyperparameters in them.","PeriodicalId":419019,"journal":{"name":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","volume":"2649 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"13","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2nd International Workshop on Embedded and Mobile Deep Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3212725.3212726","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 13

Abstract

Deep Neural Networks (DNNs) have proved to be a convenient and powerful tool for a wide range of problems. However, the extensive computational and memory resource requirements hinder the adoption of DNNs in resource-constrained scenarios. Existing compression methods have been shown to significantly reduce the computation and memory requirements of many popular DNNs. These methods, however, remain elusive to non-experts, as they demand extensive manual tuning of hyperparameters. The effects of combining various compression techniques lack exploration because of the large design space. To alleviate these challenges, this paper proposes an automated framework, Mayo, which is built on top of TensorFlow and can compress DNNs with minimal human intervention. First, we present overriders which are recursively-compositional and can be configured to effectively compress individual components (e.g. weights, biases, layer computations and gradients) in a DNN. Second, we introduce novel heuristics and a global search algorithm to efficiently optimize hyperparameters. We demonstrate that without any manual tuning, Mayo generates a sparse ResNet-18 that is 5.13x smaller than the baseline with no loss in test accuracy. By composing multiple overriders, our tool produces a sparse 6-bit CIFAR-10 classifier with only 0.16% top-1 accuracy loss and a 34x compression rate. Mayo and all compressed models are publicly available. To our knowledge, Mayo is the first framework that supports overlapping multiple compression techniques and automatically optimizes hyperparameters in them.
梅奥:一个自动生成硬件友好深度神经网络的框架
深度神经网络(dnn)已被证明是一种方便而强大的工具,可用于解决广泛的问题。然而,大量的计算和内存资源需求阻碍了dnn在资源受限情况下的采用。现有的压缩方法已经被证明可以显著减少许多流行的深度神经网络的计算和内存需求。然而,这些方法对于非专家来说仍然难以捉摸,因为它们需要大量的超参数手动调优。由于设计空间大,多种压缩技术组合的效果缺乏探索。为了缓解这些挑战,本文提出了一个自动化框架,Mayo,它建立在TensorFlow之上,可以在最少的人为干预下压缩dnn。首先,我们提出了递归组合的覆盖器,可以配置为有效地压缩DNN中的单个组件(例如权重,偏差,层计算和梯度)。其次,我们引入了新的启发式算法和全局搜索算法来有效地优化超参数。我们证明,在没有任何手动调优的情况下,Mayo生成的稀疏ResNet-18比基线小5.13倍,测试精度没有损失。通过组合多个重写器,我们的工具生成了一个稀疏的6位CIFAR-10分类器,其top-1精度损失仅为0.16%,压缩率为34倍。Mayo和所有压缩模型都是公开的。据我们所知,Mayo是第一个支持重叠多个压缩技术并自动优化其中的超参数的框架。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信