Learning the number of filters in convolutional neural networks

Jue Li, F. Cao, Honghong Cheng, Yuhua Qian
{"title":"Learning the number of filters in convolutional neural networks","authors":"Jue Li, F. Cao, Honghong Cheng, Yuhua Qian","doi":"10.1504/IJBIC.2021.114101","DOIUrl":null,"url":null,"abstract":"Convolutional networks bring the performance of many computer vision tasks to unprecedented heights, but at the cost of enormous computation load. To reduce this cost, many model compression tasks have been proposed by eliminating insignificant model structures. For example, convolution filters with small absolute weights are pruned and then fine-tuned to restore reasonable accuracy. However, most of these works rely on pre-trained models without specific analysis of the changes in filters during the training process, resulting in sizable model retraining costs. Different from previous works, we interpret the change of filter behaviour during training from the associated angle, and propose a novel filter pruning method utilising the change rule, which can remove filters with similar functions later in training. According to this strategy, not only can we achieve model compression without fine-tuning, but we can also find a novel perspective to interpret the changing behaviour of the filter during training. Moreover, our approach has been proved to be effective for many advanced CNN architectures.","PeriodicalId":13636,"journal":{"name":"Int. J. Bio Inspired Comput.","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Int. J. Bio Inspired Comput.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1504/IJBIC.2021.114101","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

Convolutional networks bring the performance of many computer vision tasks to unprecedented heights, but at the cost of enormous computation load. To reduce this cost, many model compression tasks have been proposed by eliminating insignificant model structures. For example, convolution filters with small absolute weights are pruned and then fine-tuned to restore reasonable accuracy. However, most of these works rely on pre-trained models without specific analysis of the changes in filters during the training process, resulting in sizable model retraining costs. Different from previous works, we interpret the change of filter behaviour during training from the associated angle, and propose a novel filter pruning method utilising the change rule, which can remove filters with similar functions later in training. According to this strategy, not only can we achieve model compression without fine-tuning, but we can also find a novel perspective to interpret the changing behaviour of the filter during training. Moreover, our approach has been proved to be effective for many advanced CNN architectures.
学习卷积神经网络中过滤器的数量
卷积网络使许多计算机视觉任务的性能达到了前所未有的高度,但代价是巨大的计算负荷。为了降低这一成本,许多模型压缩任务都是通过消除无关紧要的模型结构来实现的。例如,对绝对权值较小的卷积滤波器进行修剪,然后进行微调以恢复合理的精度。然而,这些工作大多依赖于预训练的模型,而没有具体分析训练过程中过滤器的变化,导致相当大的模型再训练成本。与以往的研究不同,我们从关联的角度解释了训练过程中滤波器行为的变化,并提出了一种利用变化规则的滤波器剪枝方法,该方法可以在训练后期去除具有相似功能的滤波器。根据这一策略,我们不仅可以在没有微调的情况下实现模型压缩,而且我们还可以找到一个新的角度来解释过滤器在训练过程中的变化行为。此外,我们的方法已被证明对许多先进的CNN架构是有效的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信