高效压缩:基于通道剪枝和知识蒸馏的模型压缩方法

Junan Lin, Zekai Ye, Junhan Wang
{"title":"高效压缩:基于通道剪枝和知识蒸馏的模型压缩方法","authors":"Junan Lin, Zekai Ye, Junhan Wang","doi":"10.1109/ACEDPI58926.2023.00058","DOIUrl":null,"url":null,"abstract":"With the development of neural network models in the field of deep learning image processing, more complex and higher performance neural network models can be used, but with it comes greater parameter quantities and computing power requirements. Although the model compression method has been updated in recent years, it still has problems such as inefficiency and reduced accuracy. Based on this, we propose a model compression method based on channel pruning and knowledge distillation, we sort and prune the L2 regularization weights of the convolutional layer filter containing a large number of calculations in the original model and prune it according to a certain proportion, and assist in restoring the accuracy by the knowledge distillation of the pruning model through the original model. Experimental results of the VGG16 model show that our pruning method reduced the parameters of the model by half and increased its processing speed to 2.7 times that of the original model, but the accuracy was reduced by about 10%. To solve this problem, we propose a knowledge distillation method to assist the training of the pruning model, so that the problem of model accuracy degradation has been improved and maintained at about 3.3%, to achieve a balanced state of model compression and accuracy reduction, thereby realizing the efficient compression of the model.","PeriodicalId":124469,"journal":{"name":"2023 Asia-Europe Conference on Electronics, Data Processing and Informatics (ACEDPI)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"High Efficient Compression: Model Compression Method Based on Channel Pruning and Knowledge Distillation\",\"authors\":\"Junan Lin, Zekai Ye, Junhan Wang\",\"doi\":\"10.1109/ACEDPI58926.2023.00058\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the development of neural network models in the field of deep learning image processing, more complex and higher performance neural network models can be used, but with it comes greater parameter quantities and computing power requirements. Although the model compression method has been updated in recent years, it still has problems such as inefficiency and reduced accuracy. Based on this, we propose a model compression method based on channel pruning and knowledge distillation, we sort and prune the L2 regularization weights of the convolutional layer filter containing a large number of calculations in the original model and prune it according to a certain proportion, and assist in restoring the accuracy by the knowledge distillation of the pruning model through the original model. Experimental results of the VGG16 model show that our pruning method reduced the parameters of the model by half and increased its processing speed to 2.7 times that of the original model, but the accuracy was reduced by about 10%. To solve this problem, we propose a knowledge distillation method to assist the training of the pruning model, so that the problem of model accuracy degradation has been improved and maintained at about 3.3%, to achieve a balanced state of model compression and accuracy reduction, thereby realizing the efficient compression of the model.\",\"PeriodicalId\":124469,\"journal\":{\"name\":\"2023 Asia-Europe Conference on Electronics, Data Processing and Informatics (ACEDPI)\",\"volume\":\"12 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 Asia-Europe Conference on Electronics, Data Processing and Informatics (ACEDPI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ACEDPI58926.2023.00058\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 Asia-Europe Conference on Electronics, Data Processing and Informatics (ACEDPI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACEDPI58926.2023.00058","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

随着神经网络模型在深度学习图像处理领域的发展,可以使用更复杂、性能更高的神经网络模型,但随之而来的是对参数量和计算能力的更高要求。虽然近年来模型压缩方法不断更新,但仍然存在效率低下、精度降低等问题。在此基础上,我们提出了一种基于通道剪枝和知识蒸馏的模型压缩方法,对原始模型中包含大量计算的卷积层滤波器的L2正则化权值进行排序和剪枝,并按一定比例进行剪枝,通过原始模型对剪枝模型进行知识蒸馏,辅助恢复精度。VGG16模型的实验结果表明,我们的剪枝方法使模型的参数减少了一半,处理速度提高到原模型的2.7倍,但精度降低了约10%。针对这一问题,我们提出了一种知识蒸馏的方法来辅助剪枝模型的训练,使模型精度退化问题得到改善,保持在3.3%左右,达到了模型压缩和精度降低的平衡状态,从而实现了模型的高效压缩。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
High Efficient Compression: Model Compression Method Based on Channel Pruning and Knowledge Distillation
With the development of neural network models in the field of deep learning image processing, more complex and higher performance neural network models can be used, but with it comes greater parameter quantities and computing power requirements. Although the model compression method has been updated in recent years, it still has problems such as inefficiency and reduced accuracy. Based on this, we propose a model compression method based on channel pruning and knowledge distillation, we sort and prune the L2 regularization weights of the convolutional layer filter containing a large number of calculations in the original model and prune it according to a certain proportion, and assist in restoring the accuracy by the knowledge distillation of the pruning model through the original model. Experimental results of the VGG16 model show that our pruning method reduced the parameters of the model by half and increased its processing speed to 2.7 times that of the original model, but the accuracy was reduced by about 10%. To solve this problem, we propose a knowledge distillation method to assist the training of the pruning model, so that the problem of model accuracy degradation has been improved and maintained at about 3.3%, to achieve a balanced state of model compression and accuracy reduction, thereby realizing the efficient compression of the model.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信