{"title":"深度卷积神经网络压缩的贝叶斯滤波剪枝","authors":"Haomin Lin, Tianyou Yu","doi":"10.1109/ICCECE58074.2023.10135208","DOIUrl":null,"url":null,"abstract":"Network pruning has been demonstrated as a feasible approach in reducing model complexity and accelerating the process of inference, which make it possible to deploy deep neural network in resource-limited devices. Many previous works on network pruning consider the magnitude of parameters or other intrinsic properties in point-estimates based network as the criterion of module selection, which are incapable of estimating uncertainty of parameters. In this paper, we propose a novel Bayesian filter pruning method, which leverages the advantage of Bayesian Deep Learning (BDL), by exploring the properties of distribution in weight. The proposed method removes redundant filters from a Bayesian network by a criterion of the proposed Signal to Noise Ratio (SNR) that combines properties of importance with uncertainty of filters. Experimental results on two benchmark datasets show the efficiency of our method in maintaining balance between compression and acceleration.","PeriodicalId":120030,"journal":{"name":"2023 3rd International Conference on Consumer Electronics and Computer Engineering (ICCECE)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Bayesian Filter Pruning for Deep Convolutional Neural Network Compression\",\"authors\":\"Haomin Lin, Tianyou Yu\",\"doi\":\"10.1109/ICCECE58074.2023.10135208\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Network pruning has been demonstrated as a feasible approach in reducing model complexity and accelerating the process of inference, which make it possible to deploy deep neural network in resource-limited devices. Many previous works on network pruning consider the magnitude of parameters or other intrinsic properties in point-estimates based network as the criterion of module selection, which are incapable of estimating uncertainty of parameters. In this paper, we propose a novel Bayesian filter pruning method, which leverages the advantage of Bayesian Deep Learning (BDL), by exploring the properties of distribution in weight. The proposed method removes redundant filters from a Bayesian network by a criterion of the proposed Signal to Noise Ratio (SNR) that combines properties of importance with uncertainty of filters. Experimental results on two benchmark datasets show the efficiency of our method in maintaining balance between compression and acceleration.\",\"PeriodicalId\":120030,\"journal\":{\"name\":\"2023 3rd International Conference on Consumer Electronics and Computer Engineering (ICCECE)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 3rd International Conference on Consumer Electronics and Computer Engineering (ICCECE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCECE58074.2023.10135208\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 3rd International Conference on Consumer Electronics and Computer Engineering (ICCECE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCECE58074.2023.10135208","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Bayesian Filter Pruning for Deep Convolutional Neural Network Compression
Network pruning has been demonstrated as a feasible approach in reducing model complexity and accelerating the process of inference, which make it possible to deploy deep neural network in resource-limited devices. Many previous works on network pruning consider the magnitude of parameters or other intrinsic properties in point-estimates based network as the criterion of module selection, which are incapable of estimating uncertainty of parameters. In this paper, we propose a novel Bayesian filter pruning method, which leverages the advantage of Bayesian Deep Learning (BDL), by exploring the properties of distribution in weight. The proposed method removes redundant filters from a Bayesian network by a criterion of the proposed Signal to Noise Ratio (SNR) that combines properties of importance with uncertainty of filters. Experimental results on two benchmark datasets show the efficiency of our method in maintaining balance between compression and acceleration.