Efficient deep convolutional model compression with an active stepwise pruning approach

Sheng-sheng Wang, Chunshang Xing, Dong Liu
{"title":"Efficient deep convolutional model compression with an active stepwise pruning approach","authors":"Sheng-sheng Wang, Chunshang Xing, Dong Liu","doi":"10.1504/ijcse.2020.10031600","DOIUrl":null,"url":null,"abstract":"Deep models are structurally tremendous and complex, thus making it hard to deploy on the embedded hardware with restricted memory and computing power. Although, the existing compression methods have pruned the deep models effectively, some issues exist in those methods, such as multiple iterations needed in fine-tuning phase, difficulty in pruning granularity control and numerous hyperparameters needed to set. In this paper, we propose an active stepwise pruning method of a logarithmic function which only needs to set three hyperparameters and a few epochs. We also propose a recovery strategy to repair the incorrect pruning thus ensuring the prediction accuracy of model. Pruning and repairing alternately constitute cyclic process along with updating the weights in layers. Our method can prune the parameters of MobileNet, AlexNet, VGG-16 and ZFNet by a factor of 5.6×, 11.7×, 16.6× and 15× respectively without any accuracy loss, which precedes the existing methods.","PeriodicalId":340410,"journal":{"name":"Int. J. Comput. Sci. Eng.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Int. J. Comput. Sci. Eng.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1504/ijcse.2020.10031600","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Deep models are structurally tremendous and complex, thus making it hard to deploy on the embedded hardware with restricted memory and computing power. Although, the existing compression methods have pruned the deep models effectively, some issues exist in those methods, such as multiple iterations needed in fine-tuning phase, difficulty in pruning granularity control and numerous hyperparameters needed to set. In this paper, we propose an active stepwise pruning method of a logarithmic function which only needs to set three hyperparameters and a few epochs. We also propose a recovery strategy to repair the incorrect pruning thus ensuring the prediction accuracy of model. Pruning and repairing alternately constitute cyclic process along with updating the weights in layers. Our method can prune the parameters of MobileNet, AlexNet, VGG-16 and ZFNet by a factor of 5.6×, 11.7×, 16.6× and 15× respectively without any accuracy loss, which precedes the existing methods.
有效的深度卷积模型压缩与主动逐步修剪方法
深度模型结构庞大且复杂,难以在内存和计算能力有限的嵌入式硬件上部署。现有的压缩方法虽然对深度模型进行了有效的剪枝,但存在微调阶段需要多次迭代、剪枝粒度控制困难以及需要设置大量超参数等问题。本文提出了一种对数函数的主动逐步剪枝方法,该方法只需要设置三个超参数和几个epoch。我们还提出了一种修复错误剪枝的恢复策略,从而保证了模型的预测精度。随着层间权值的更新,剪枝和修复交替构成循环过程。我们的方法可以对MobileNet、AlexNet、VGG-16和ZFNet的参数分别进行5.6倍、11.7倍、16.6倍和15倍的裁剪,而没有任何精度损失,优于现有方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信