加权压缩友好二值化神经网络

Yuzhong Jiao, Xiao Huo, Yuan Lei, Sha Li, Yiu Kei Li
{"title":"加权压缩友好二值化神经网络","authors":"Yuzhong Jiao, Xiao Huo, Yuan Lei, Sha Li, Yiu Kei Li","doi":"10.1109/GCAIoT51063.2020.9345815","DOIUrl":null,"url":null,"abstract":"The resources of edge devices in AIoT systems are usually constrained with size and power. The computational complexity of neural network models in these edge devices has become a major concern. The most compact form of deep neural networks is binarized neural network (BNN), which adopts binary weights and exclusive NOR (XNOR) operations as binary convolution. In this paper, we propose weight compression-friendly BNN to save hardware resources by reducing memory space. The proposed technique does not binarize weights just according to the signs of weights, but fully considers compression efficiency in the training of the BNN model. The experiments are performed by using the binary version of a 6-layer convolutional neural network (CNN) and MNIST case. The results show that the proposed technique can achieve more than 25% reduction in memory space with accuracy loss of 1 %, or more than 35% memory reduction with about 2.5% accuracy drop for MNIST classification. The weight compression method does not destroy the regular structure of neural networks, so the proposed technique is very fit for processor-based BNN hardware accelerators.","PeriodicalId":398815,"journal":{"name":"2020 IEEE Global Conference on Artificial Intelligence and Internet of Things (GCAIoT)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Weight Compression-Friendly Binarized Neural Network\",\"authors\":\"Yuzhong Jiao, Xiao Huo, Yuan Lei, Sha Li, Yiu Kei Li\",\"doi\":\"10.1109/GCAIoT51063.2020.9345815\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The resources of edge devices in AIoT systems are usually constrained with size and power. The computational complexity of neural network models in these edge devices has become a major concern. The most compact form of deep neural networks is binarized neural network (BNN), which adopts binary weights and exclusive NOR (XNOR) operations as binary convolution. In this paper, we propose weight compression-friendly BNN to save hardware resources by reducing memory space. The proposed technique does not binarize weights just according to the signs of weights, but fully considers compression efficiency in the training of the BNN model. The experiments are performed by using the binary version of a 6-layer convolutional neural network (CNN) and MNIST case. The results show that the proposed technique can achieve more than 25% reduction in memory space with accuracy loss of 1 %, or more than 35% memory reduction with about 2.5% accuracy drop for MNIST classification. The weight compression method does not destroy the regular structure of neural networks, so the proposed technique is very fit for processor-based BNN hardware accelerators.\",\"PeriodicalId\":398815,\"journal\":{\"name\":\"2020 IEEE Global Conference on Artificial Intelligence and Internet of Things (GCAIoT)\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-12-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE Global Conference on Artificial Intelligence and Internet of Things (GCAIoT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/GCAIoT51063.2020.9345815\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Global Conference on Artificial Intelligence and Internet of Things (GCAIoT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GCAIoT51063.2020.9345815","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

摘要

AIoT系统中边缘设备的资源通常受到尺寸和功率的限制。在这些边缘设备中,神经网络模型的计算复杂度已经成为人们关注的主要问题。深度神经网络中最紧凑的形式是二值化神经网络(BNN),它采用二元权值和异NOR (XNOR)运算作为二值卷积。在本文中,我们提出了权重压缩友好的BNN,通过减少内存空间来节省硬件资源。提出的方法并不是仅仅根据权值的符号对权值进行二值化,而是在训练BNN模型时充分考虑了压缩效率。实验采用六层卷积神经网络(CNN)的二进制版本和MNIST案例进行。结果表明,对于MNIST分类,该技术可以在精度下降1%的情况下减少25%以上的内存空间,或者在精度下降2.5%的情况下减少35%以上的内存空间。权重压缩方法不破坏神经网络的规则结构,因此该方法非常适合基于处理器的BNN硬件加速器。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Weight Compression-Friendly Binarized Neural Network
The resources of edge devices in AIoT systems are usually constrained with size and power. The computational complexity of neural network models in these edge devices has become a major concern. The most compact form of deep neural networks is binarized neural network (BNN), which adopts binary weights and exclusive NOR (XNOR) operations as binary convolution. In this paper, we propose weight compression-friendly BNN to save hardware resources by reducing memory space. The proposed technique does not binarize weights just according to the signs of weights, but fully considers compression efficiency in the training of the BNN model. The experiments are performed by using the binary version of a 6-layer convolutional neural network (CNN) and MNIST case. The results show that the proposed technique can achieve more than 25% reduction in memory space with accuracy loss of 1 %, or more than 35% memory reduction with about 2.5% accuracy drop for MNIST classification. The weight compression method does not destroy the regular structure of neural networks, so the proposed technique is very fit for processor-based BNN hardware accelerators.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信