Automatic Optimization of super Parameters Based on Model Pruning and Knowledge Distillation

Min Wu, Weihua Ma, Yue Li, Xiongbo Zhao
{"title":"Automatic Optimization of super Parameters Based on Model Pruning and Knowledge Distillation","authors":"Min Wu, Weihua Ma, Yue Li, Xiongbo Zhao","doi":"10.1109/ICCEIC51584.2020.00030","DOIUrl":null,"url":null,"abstract":"In recent years, deep neural network has been widely used in computer vision, speech recognition and other fields. However, to obtain better performance, it needs to design a network with higher complexity, and the corresponding model calculation amount and storage space are also increasing. At the same time, the computing resources and energy consumption budget of mobile devices are very limited. Therefore, model compression is very important for deploying neural network models on mobile devices. Knowledge distillation technology based on transfer learning is an effective method to realize model compression. This study proposes: the model pruning technology is introduced into the student network design of knowledge distillation, and the super parameters (temperature T, scale factor λ, pruning rate ϒ) are automatically optimized, and the optimal combination of parameters is selected as the final value according to the final performance. The results show that, compared with the commonly used pruning techniques, this method can effectively improve the accuracy of the network without increasing the network size, and the network performance can be further improved by adjusting the value of super parameters.","PeriodicalId":135840,"journal":{"name":"2020 International Conference on Computer Engineering and Intelligent Control (ICCEIC)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Computer Engineering and Intelligent Control (ICCEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCEIC51584.2020.00030","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In recent years, deep neural network has been widely used in computer vision, speech recognition and other fields. However, to obtain better performance, it needs to design a network with higher complexity, and the corresponding model calculation amount and storage space are also increasing. At the same time, the computing resources and energy consumption budget of mobile devices are very limited. Therefore, model compression is very important for deploying neural network models on mobile devices. Knowledge distillation technology based on transfer learning is an effective method to realize model compression. This study proposes: the model pruning technology is introduced into the student network design of knowledge distillation, and the super parameters (temperature T, scale factor λ, pruning rate ϒ) are automatically optimized, and the optimal combination of parameters is selected as the final value according to the final performance. The results show that, compared with the commonly used pruning techniques, this method can effectively improve the accuracy of the network without increasing the network size, and the network performance can be further improved by adjusting the value of super parameters.
基于模型剪枝和知识蒸馏的超参数自动优化
近年来,深度神经网络在计算机视觉、语音识别等领域得到了广泛应用。但是,为了获得更好的性能,需要设计复杂度更高的网络,相应的模型计算量和存储空间也在不断增加。同时,移动设备的计算资源和能耗预算非常有限。因此,模型压缩对于在移动设备上部署神经网络模型非常重要。基于迁移学习的知识蒸馏技术是实现模型压缩的有效方法。本研究提出:将模型剪枝技术引入知识蒸馏的学生网络设计中,自动优化超参数(温度T、尺度因子λ、剪枝率γ),并根据最终性能选择参数的最优组合作为最终值。结果表明,与常用的剪枝技术相比,该方法可以在不增加网络规模的情况下有效提高网络的准确率,并且通过调整超参数的值可以进一步提高网络性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信