Amplitude Suppression and Direction Activation in Networks for 1-bit Faster R-CNN

Sheng Xu, Zhendong Liu, Xuan Gong, Chunlei Liu, Mingyuan Mao, Baochang Zhang
{"title":"Amplitude Suppression and Direction Activation in Networks for 1-bit Faster R-CNN","authors":"Sheng Xu, Zhendong Liu, Xuan Gong, Chunlei Liu, Mingyuan Mao, Baochang Zhang","doi":"10.1145/3410338.3412340","DOIUrl":null,"url":null,"abstract":"Recent advances in object detection have been driven by the success of deep convolutional neural networks (DCNNs). Deploying a DCNN detector on resource-limited hardware such as embedded devices and smart phones, however, remains challenging due to the massive number of parameters a typical model contains. In this paper, we propose an amplitude suppression and direction activation for Faster R-CNN (ASDA-FRCNN) framework to significantly compress DCNNs for highly efficient performance. The shared amplitude between the full-precision and the binary kernels can be significantly suppressed through a simple but effective loss, which is then incorporated into the existing Faster R-CNN detector. Furthermore, the ASDA module is generic and flexible to be incorporated into existing DCNNs for different tasks. Experiments demonstrate the superiority of 1-bit ASDA-FRCNN which achieves superior performance on various datasets. Specifically, ASDA-FRCNN shows the best speed-accuracy trade off with 63.4% at estimated 711 FPS and 19.4% mAP at and estimated 362 FPS with ResNet-18 on the PASCAL VOC 2007 and MS COCO validation datasets respectively, which demonstrate the superior performance and strong generalization of our method.","PeriodicalId":401260,"journal":{"name":"Proceedings of the 4th International Workshop on Embedded and Mobile Deep Learning","volume":"43 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 4th International Workshop on Embedded and Mobile Deep Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3410338.3412340","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

Recent advances in object detection have been driven by the success of deep convolutional neural networks (DCNNs). Deploying a DCNN detector on resource-limited hardware such as embedded devices and smart phones, however, remains challenging due to the massive number of parameters a typical model contains. In this paper, we propose an amplitude suppression and direction activation for Faster R-CNN (ASDA-FRCNN) framework to significantly compress DCNNs for highly efficient performance. The shared amplitude between the full-precision and the binary kernels can be significantly suppressed through a simple but effective loss, which is then incorporated into the existing Faster R-CNN detector. Furthermore, the ASDA module is generic and flexible to be incorporated into existing DCNNs for different tasks. Experiments demonstrate the superiority of 1-bit ASDA-FRCNN which achieves superior performance on various datasets. Specifically, ASDA-FRCNN shows the best speed-accuracy trade off with 63.4% at estimated 711 FPS and 19.4% mAP at and estimated 362 FPS with ResNet-18 on the PASCAL VOC 2007 and MS COCO validation datasets respectively, which demonstrate the superior performance and strong generalization of our method.
1位更快R-CNN网络的幅度抑制和方向激活
深度卷积神经网络(DCNNs)的成功推动了目标检测的最新进展。然而,在资源有限的硬件(如嵌入式设备和智能手机)上部署DCNN检测器仍然具有挑战性,因为典型模型包含大量参数。在本文中,我们提出了一种用于Faster R-CNN (ASDA-FRCNN)框架的幅度抑制和方向激活,以显著压缩DCNNs以获得高效的性能。全精度核和二值核之间的共享幅度可以通过简单而有效的损失来显著抑制,然后将其纳入现有的Faster R-CNN检测器中。此外,ASDA模块具有通用性和灵活性,可以将其集成到现有的DCNNs中以执行不同的任务。实验证明了1位ASDA-FRCNN的优越性,在各种数据集上都取得了优异的性能。具体而言,在PASCAL VOC 2007和MS COCO验证数据集上,ASDA-FRCNN显示出最佳的速度-精度权衡,在估计711 FPS时,ASDA-FRCNN达到63.4%,在ResNet-18下,mAP和362 FPS分别达到19.4%和19.4%,这表明我们的方法具有优越的性能和强大的泛化性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信