{"title":"基于模型剪枝和知识蒸馏的超参数自动优化","authors":"Min Wu, Weihua Ma, Yue Li, Xiongbo Zhao","doi":"10.1109/ICCEIC51584.2020.00030","DOIUrl":null,"url":null,"abstract":"In recent years, deep neural network has been widely used in computer vision, speech recognition and other fields. However, to obtain better performance, it needs to design a network with higher complexity, and the corresponding model calculation amount and storage space are also increasing. At the same time, the computing resources and energy consumption budget of mobile devices are very limited. Therefore, model compression is very important for deploying neural network models on mobile devices. Knowledge distillation technology based on transfer learning is an effective method to realize model compression. This study proposes: the model pruning technology is introduced into the student network design of knowledge distillation, and the super parameters (temperature T, scale factor λ, pruning rate ϒ) are automatically optimized, and the optimal combination of parameters is selected as the final value according to the final performance. The results show that, compared with the commonly used pruning techniques, this method can effectively improve the accuracy of the network without increasing the network size, and the network performance can be further improved by adjusting the value of super parameters.","PeriodicalId":135840,"journal":{"name":"2020 International Conference on Computer Engineering and Intelligent Control (ICCEIC)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Automatic Optimization of super Parameters Based on Model Pruning and Knowledge Distillation\",\"authors\":\"Min Wu, Weihua Ma, Yue Li, Xiongbo Zhao\",\"doi\":\"10.1109/ICCEIC51584.2020.00030\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, deep neural network has been widely used in computer vision, speech recognition and other fields. However, to obtain better performance, it needs to design a network with higher complexity, and the corresponding model calculation amount and storage space are also increasing. At the same time, the computing resources and energy consumption budget of mobile devices are very limited. Therefore, model compression is very important for deploying neural network models on mobile devices. Knowledge distillation technology based on transfer learning is an effective method to realize model compression. This study proposes: the model pruning technology is introduced into the student network design of knowledge distillation, and the super parameters (temperature T, scale factor λ, pruning rate ϒ) are automatically optimized, and the optimal combination of parameters is selected as the final value according to the final performance. The results show that, compared with the commonly used pruning techniques, this method can effectively improve the accuracy of the network without increasing the network size, and the network performance can be further improved by adjusting the value of super parameters.\",\"PeriodicalId\":135840,\"journal\":{\"name\":\"2020 International Conference on Computer Engineering and Intelligent Control (ICCEIC)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 International Conference on Computer Engineering and Intelligent Control (ICCEIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCEIC51584.2020.00030\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Computer Engineering and Intelligent Control (ICCEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCEIC51584.2020.00030","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Automatic Optimization of super Parameters Based on Model Pruning and Knowledge Distillation
In recent years, deep neural network has been widely used in computer vision, speech recognition and other fields. However, to obtain better performance, it needs to design a network with higher complexity, and the corresponding model calculation amount and storage space are also increasing. At the same time, the computing resources and energy consumption budget of mobile devices are very limited. Therefore, model compression is very important for deploying neural network models on mobile devices. Knowledge distillation technology based on transfer learning is an effective method to realize model compression. This study proposes: the model pruning technology is introduced into the student network design of knowledge distillation, and the super parameters (temperature T, scale factor λ, pruning rate ϒ) are automatically optimized, and the optimal combination of parameters is selected as the final value according to the final performance. The results show that, compared with the commonly used pruning techniques, this method can effectively improve the accuracy of the network without increasing the network size, and the network performance can be further improved by adjusting the value of super parameters.