{"title":"A Summary of convolution Neural Network Compression and Acceleration Technology","authors":"Bingzhen Li, Wenzhi Jiang, Jiaojiao Gu, Ke Liu","doi":"10.1109/ICHCI51889.2020.00065","DOIUrl":null,"url":null,"abstract":"Although convolution neural network has achieved remarkable results in different application scenarios, there are a large number of parameters and computation in its structure, which limit its development in mobile and embedded devices. How to reduce parameters, compress model and optimize structure to improve network performance without losing accuracy has become a hot issue of convolution neural network. This paper summarizes and summarizes the convolution neural network structure optimization technology from five aspects: granularity pruning, weight quantization sharing, knowledge distillation, tensor decomposition and fine network design, and analyzes the technical core of it. Their advantages and disadvantages, applicable scenarios and optimization results are analyzed and summarized respectively, and the future research direction is prospected.","PeriodicalId":355427,"journal":{"name":"2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICHCI51889.2020.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Although convolution neural network has achieved remarkable results in different application scenarios, there are a large number of parameters and computation in its structure, which limit its development in mobile and embedded devices. How to reduce parameters, compress model and optimize structure to improve network performance without losing accuracy has become a hot issue of convolution neural network. This paper summarizes and summarizes the convolution neural network structure optimization technology from five aspects: granularity pruning, weight quantization sharing, knowledge distillation, tensor decomposition and fine network design, and analyzes the technical core of it. Their advantages and disadvantages, applicable scenarios and optimization results are analyzed and summarized respectively, and the future research direction is prospected.