{"title":"Speeding Up Deep Convolutional Neural Networks Based on Tucker-CP Decomposition","authors":"Dechun Song, Peiyong Zhang, Feiteng Li","doi":"10.1145/3409073.3409094","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks (CNNs) have made great success in computer vision tasks. But the computational complexity of CNNs is huge, which makes CNNs run slowly especially when computational resources are limited. In this paper, we propose a scheme based on tensor decomposition to accelerate CNNs. Firstly, Tucker method is used to decompose the convolution kernel into a small core tensor with key information and two factor matrices reflecting the linear relationship in the third dimension and fourth dimension of the convolution kernel respectively. Then CP (CANDECOMP/PARAFAC) method is used to decompose the core tensor into several rank-1 tensors. This scheme can remove the linear redundancy in convolution kernels and greatly speed up CNNs while maintaining the high classification accuracy. The scheme is used to decompose all the convolutional layers in AlexNet, and the accelerated model is trained and tested on ImageNet. The results show that our scheme achieves a whole-model speedup of 4 x with merely a 1.9% increase in top-5 error for AlexNet.","PeriodicalId":229746,"journal":{"name":"Proceedings of the 2020 5th International Conference on Machine Learning Technologies","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 5th International Conference on Machine Learning Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3409073.3409094","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6
Abstract
Convolutional neural networks (CNNs) have made great success in computer vision tasks. But the computational complexity of CNNs is huge, which makes CNNs run slowly especially when computational resources are limited. In this paper, we propose a scheme based on tensor decomposition to accelerate CNNs. Firstly, Tucker method is used to decompose the convolution kernel into a small core tensor with key information and two factor matrices reflecting the linear relationship in the third dimension and fourth dimension of the convolution kernel respectively. Then CP (CANDECOMP/PARAFAC) method is used to decompose the core tensor into several rank-1 tensors. This scheme can remove the linear redundancy in convolution kernels and greatly speed up CNNs while maintaining the high classification accuracy. The scheme is used to decompose all the convolutional layers in AlexNet, and the accelerated model is trained and tested on ImageNet. The results show that our scheme achieves a whole-model speedup of 4 x with merely a 1.9% increase in top-5 error for AlexNet.