{"title":"基于权重分布的k均值深度卷积网络压缩","authors":"Wang Lei, Huawei Chen, Yixuan Wu","doi":"10.1145/3144789.3144803","DOIUrl":null,"url":null,"abstract":"For the application of deep neural networks on devices with limited hardware resources, it is necessary to reduce the computational complexity and storage requirement. Compression is an effective way to achieve this goal. One of the available method is to quantize the weights to enforce weight sharing, which can greatly reduce the parameters of each layer. This paper presents an improved k-means clustering algorithm to compress CNN (convolutional neural networks).By taking weights distribution into consideration when choosing clustering centers to quantize weights, this algorithm automatically chooses and revises centers to compress network. Compared with traditional quantification method, this algorithm can maintain accuracy and increase the compression speed at the same time. Experiments on AlexNet show that using k-means based on weights distribution to quantize the weights can improve compression speed by 5% to 10% and improve accuracy by 6% compared to traditional algorithm. This algorithm provides a better way for the application of convolutional neural networks on mobile devices.","PeriodicalId":254163,"journal":{"name":"Proceedings of the 2nd International Conference on Intelligent Information Processing","volume":"86 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Compressing Deep Convolutional Networks Using K-means Based on Weights Distribution\",\"authors\":\"Wang Lei, Huawei Chen, Yixuan Wu\",\"doi\":\"10.1145/3144789.3144803\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"For the application of deep neural networks on devices with limited hardware resources, it is necessary to reduce the computational complexity and storage requirement. Compression is an effective way to achieve this goal. One of the available method is to quantize the weights to enforce weight sharing, which can greatly reduce the parameters of each layer. This paper presents an improved k-means clustering algorithm to compress CNN (convolutional neural networks).By taking weights distribution into consideration when choosing clustering centers to quantize weights, this algorithm automatically chooses and revises centers to compress network. Compared with traditional quantification method, this algorithm can maintain accuracy and increase the compression speed at the same time. Experiments on AlexNet show that using k-means based on weights distribution to quantize the weights can improve compression speed by 5% to 10% and improve accuracy by 6% compared to traditional algorithm. This algorithm provides a better way for the application of convolutional neural networks on mobile devices.\",\"PeriodicalId\":254163,\"journal\":{\"name\":\"Proceedings of the 2nd International Conference on Intelligent Information Processing\",\"volume\":\"86 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-07-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2nd International Conference on Intelligent Information Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3144789.3144803\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2nd International Conference on Intelligent Information Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3144789.3144803","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Compressing Deep Convolutional Networks Using K-means Based on Weights Distribution
For the application of deep neural networks on devices with limited hardware resources, it is necessary to reduce the computational complexity and storage requirement. Compression is an effective way to achieve this goal. One of the available method is to quantize the weights to enforce weight sharing, which can greatly reduce the parameters of each layer. This paper presents an improved k-means clustering algorithm to compress CNN (convolutional neural networks).By taking weights distribution into consideration when choosing clustering centers to quantize weights, this algorithm automatically chooses and revises centers to compress network. Compared with traditional quantification method, this algorithm can maintain accuracy and increase the compression speed at the same time. Experiments on AlexNet show that using k-means based on weights distribution to quantize the weights can improve compression speed by 5% to 10% and improve accuracy by 6% compared to traditional algorithm. This algorithm provides a better way for the application of convolutional neural networks on mobile devices.