{"title":"按层组的重要程度进行滤波剪枝,以加速卷积神经网络","authors":"Yunseok Jang, Jaeseok Kim","doi":"10.35803/1694-5298.2021.3.358-365","DOIUrl":null,"url":null,"abstract":"Various acceleration approaches have been studied to deploy convolutional neural networks in embedded devices. Among them, filter pruning is the most active research because it is easy to implement in hardware and keeps high accuracy while reducing the computational and memory cost. In this paper, we propose a method of grouping layers, finding the importance of each group, and groupwise pruning according to the order of importance to achieve high FLOPs reduction while retaining high accuracy. First, we divide the layers of the pre-trained network into groups according to the size of the output feature map. Next, we calculate the importance score per group using first-order Taylor expansion. Finally, filter pruning is performed in order from the group with the highest importance score. When pruning VGG and ResNet trained on CIFAR-10, our proposed method shows superior performance in accuracy and FLOPs compared to the state-of-art methods. Notably, on ResNet-50, we achieve 70.85% FLOPs reduction by removing 50% of the filters, with a slight loss of 0.41% in the baseline accuracy.","PeriodicalId":22490,"journal":{"name":"The Herald of KSUCTA, №3, 2021","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FILTER PRUNING IN ORDER OF IMPORTANCE OF THE LAYER GROUPS TO ACCELERATE CONVOLUTIONAL NEURAL NETWORKS\",\"authors\":\"Yunseok Jang, Jaeseok Kim\",\"doi\":\"10.35803/1694-5298.2021.3.358-365\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Various acceleration approaches have been studied to deploy convolutional neural networks in embedded devices. Among them, filter pruning is the most active research because it is easy to implement in hardware and keeps high accuracy while reducing the computational and memory cost. In this paper, we propose a method of grouping layers, finding the importance of each group, and groupwise pruning according to the order of importance to achieve high FLOPs reduction while retaining high accuracy. First, we divide the layers of the pre-trained network into groups according to the size of the output feature map. Next, we calculate the importance score per group using first-order Taylor expansion. Finally, filter pruning is performed in order from the group with the highest importance score. When pruning VGG and ResNet trained on CIFAR-10, our proposed method shows superior performance in accuracy and FLOPs compared to the state-of-art methods. Notably, on ResNet-50, we achieve 70.85% FLOPs reduction by removing 50% of the filters, with a slight loss of 0.41% in the baseline accuracy.\",\"PeriodicalId\":22490,\"journal\":{\"name\":\"The Herald of KSUCTA, №3, 2021\",\"volume\":\"1 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-09-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Herald of KSUCTA, №3, 2021\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.35803/1694-5298.2021.3.358-365\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Herald of KSUCTA, №3, 2021","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.35803/1694-5298.2021.3.358-365","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
FILTER PRUNING IN ORDER OF IMPORTANCE OF THE LAYER GROUPS TO ACCELERATE CONVOLUTIONAL NEURAL NETWORKS
Various acceleration approaches have been studied to deploy convolutional neural networks in embedded devices. Among them, filter pruning is the most active research because it is easy to implement in hardware and keeps high accuracy while reducing the computational and memory cost. In this paper, we propose a method of grouping layers, finding the importance of each group, and groupwise pruning according to the order of importance to achieve high FLOPs reduction while retaining high accuracy. First, we divide the layers of the pre-trained network into groups according to the size of the output feature map. Next, we calculate the importance score per group using first-order Taylor expansion. Finally, filter pruning is performed in order from the group with the highest importance score. When pruning VGG and ResNet trained on CIFAR-10, our proposed method shows superior performance in accuracy and FLOPs compared to the state-of-art methods. Notably, on ResNet-50, we achieve 70.85% FLOPs reduction by removing 50% of the filters, with a slight loss of 0.41% in the baseline accuracy.