Yingjian Ling, Kan Zhong, Yunsong Wu, Duo Liu, Jinting Ren, Renping Liu, Moming Duan, Weichen Liu, Liang Liang
{"title":"面向嵌入式系统的部分二值化卷积神经网络","authors":"Yingjian Ling, Kan Zhong, Yunsong Wu, Duo Liu, Jinting Ren, Renping Liu, Moming Duan, Weichen Liu, Liang Liang","doi":"10.1109/ISVLSI.2018.00034","DOIUrl":null,"url":null,"abstract":"We have witnessed the tremendous success of deep neural networks. However, this success comes with the considerable computation and storage costs which make it difficult to deploy these networks directly on resource-constrained embedded systems. To address this problem, we propose TaiJiNet, a binary-network-based framework that combines binary convolutions and pointwise convolutions, to reduce the computation and storage overhead while maintaining a comparable accuracy. Furthermore, in order to provide TaiJiNet with more flexibility, we introduce a strategy called partial binarized convolution to efficiently balance network performance and accuracy. We evaluate TaiJiNet with the CIFAR-10 and ImageNet datasets. The experimental results show that with the proposed TaiJiNet framework, the binary version of AlexNet can achieve 26x compression rate with a negligible 0.8% accuracy drop when compared with the full-precision AlexNet.","PeriodicalId":114330,"journal":{"name":"2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"TaiJiNet: Towards Partial Binarized Convolutional Neural Network for Embedded Systems\",\"authors\":\"Yingjian Ling, Kan Zhong, Yunsong Wu, Duo Liu, Jinting Ren, Renping Liu, Moming Duan, Weichen Liu, Liang Liang\",\"doi\":\"10.1109/ISVLSI.2018.00034\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We have witnessed the tremendous success of deep neural networks. However, this success comes with the considerable computation and storage costs which make it difficult to deploy these networks directly on resource-constrained embedded systems. To address this problem, we propose TaiJiNet, a binary-network-based framework that combines binary convolutions and pointwise convolutions, to reduce the computation and storage overhead while maintaining a comparable accuracy. Furthermore, in order to provide TaiJiNet with more flexibility, we introduce a strategy called partial binarized convolution to efficiently balance network performance and accuracy. We evaluate TaiJiNet with the CIFAR-10 and ImageNet datasets. The experimental results show that with the proposed TaiJiNet framework, the binary version of AlexNet can achieve 26x compression rate with a negligible 0.8% accuracy drop when compared with the full-precision AlexNet.\",\"PeriodicalId\":114330,\"journal\":{\"name\":\"2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)\",\"volume\":\"20 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISVLSI.2018.00034\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISVLSI.2018.00034","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
TaiJiNet: Towards Partial Binarized Convolutional Neural Network for Embedded Systems
We have witnessed the tremendous success of deep neural networks. However, this success comes with the considerable computation and storage costs which make it difficult to deploy these networks directly on resource-constrained embedded systems. To address this problem, we propose TaiJiNet, a binary-network-based framework that combines binary convolutions and pointwise convolutions, to reduce the computation and storage overhead while maintaining a comparable accuracy. Furthermore, in order to provide TaiJiNet with more flexibility, we introduce a strategy called partial binarized convolution to efficiently balance network performance and accuracy. We evaluate TaiJiNet with the CIFAR-10 and ImageNet datasets. The experimental results show that with the proposed TaiJiNet framework, the binary version of AlexNet can achieve 26x compression rate with a negligible 0.8% accuracy drop when compared with the full-precision AlexNet.