An Evolution Approach for Pre-trained Neural Network Pruning without Original Training Dataset

Toan Pham Van, T. Tung, Linh Bao Doan, Thanh Ta Minh
{"title":"An Evolution Approach for Pre-trained Neural Network Pruning without Original Training Dataset","authors":"Toan Pham Van, T. Tung, Linh Bao Doan, Thanh Ta Minh","doi":"10.18178/ijke.2022.8.1.136","DOIUrl":null,"url":null,"abstract":"—Model pruning is an important technique in real-world machine learning problems, especially in deep learning. This technique has provided some methods for compressing a large model to a smaller model while retaining the most accuracy. However, a majority of these approaches require a full original training set. This might not always be possible in practice if the model is trained in a large-scale dataset or on a dataset whose release poses privacy. Although we cannot access the original training set in some cases, pre-trained models are available more often. This paper aims to solve the model pruning problem without the initial training set by finding the sub-networks in the initial pre-trained model. We propose an approach of using genetic algorithms (GA) to find the sub-networks systematically and automatically. Experimental results show that our algorithm can find good sub-networks efficiently. Theoretically, if we had unlimited time and hardware power, we could find the optimized sub-networks of any pre-trained model and achieve the best results in the future. Our code and pre-trained models are available at: https://github.com/sun-asterisk-research/ga_pruning_research.","PeriodicalId":88527,"journal":{"name":"International journal of knowledge engineering and soft data paradigms","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International journal of knowledge engineering and soft data paradigms","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.18178/ijke.2022.8.1.136","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

—Model pruning is an important technique in real-world machine learning problems, especially in deep learning. This technique has provided some methods for compressing a large model to a smaller model while retaining the most accuracy. However, a majority of these approaches require a full original training set. This might not always be possible in practice if the model is trained in a large-scale dataset or on a dataset whose release poses privacy. Although we cannot access the original training set in some cases, pre-trained models are available more often. This paper aims to solve the model pruning problem without the initial training set by finding the sub-networks in the initial pre-trained model. We propose an approach of using genetic algorithms (GA) to find the sub-networks systematically and automatically. Experimental results show that our algorithm can find good sub-networks efficiently. Theoretically, if we had unlimited time and hardware power, we could find the optimized sub-networks of any pre-trained model and achieve the best results in the future. Our code and pre-trained models are available at: https://github.com/sun-asterisk-research/ga_pruning_research.
一种无原始训练数据集的神经网络预训练剪枝进化方法
模型修剪是现实世界机器学习问题中的一项重要技术,特别是在深度学习中。该技术提供了一些将大模型压缩到小模型的方法,同时保持了最大的准确性。然而,这些方法中的大多数都需要一个完整的原始训练集。如果模型是在大规模数据集中训练的,或者在发布时会带来隐私的数据集上训练,那么在实践中这可能并不总是可能的。虽然在某些情况下我们无法访问原始训练集,但预训练模型更常见。本文旨在通过在初始预训练模型中寻找子网络来解决没有初始训练集的模型剪枝问题。提出了一种利用遗传算法系统地、自动地寻找子网络的方法。实验结果表明,该算法能有效地找到较好的子网络。理论上,如果我们有无限的时间和硬件能力,我们可以找到任何预训练模型的优化子网络,并在未来获得最佳结果。我们的代码和预训练模型可在:https://github.com/sun-asterisk-research/ga_pruning_research。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信