使用无监督深度学习结构进行异常检测的快速硬件辅助在线学习

Khaled Alrawashdeh, C. Purdy
{"title":"使用无监督深度学习结构进行异常检测的快速硬件辅助在线学习","authors":"Khaled Alrawashdeh, C. Purdy","doi":"10.1109/INFOCT.2018.8356855","DOIUrl":null,"url":null,"abstract":"Real-time scenarios of deep learning algorithms are challenged by two less frequently addressed issues. The first is data inefficiency i.e., the model requires several epochs of trial and error to converge which makes it impractical to be applied to real-time applications. The Second is the high precision computation load of the deep learning algorithms to achieve high accuracy during training and inference. In this paper, we address these two issues and apply our model to the task of online anomaly detection using FPGA. To address the first issue, we propose a compressed training model for the contrastive divergence algorithm (CD) in the Deep Belief Network (DBN). The goal is to dynamically adjust the training vector according to the feedback from the free energy and the reconstruction error, which allows for better generalization. To address the second issue, we propose a Hybrid-Stochastic-Dynamic-Fixed-Point (HSDFP) method, which provides training environment with high reduction in calculation, area, and power in FPGA. Our framework enables the DBN structure to take actions and detect attacks online. Thus, the network can collect efficient number of training samples and avoid overfitting. We show that (1) our proposed method converges faster than the state-of-the-art deep learning methods, (2) FPGA implementation achieves accelerated inference speed of 0.008ms and a high power efficiency of 37 G-ops/s/W compared to CPU, GPU, and 16-bit fixed-point arithmetic (3) FPGA also achieves minimal degradation in accuracy of 95%, 95.4%, and 97.9% on the benchmark datasets: MNIST, NSL-KDD, and Kyoto datasets.","PeriodicalId":376443,"journal":{"name":"2018 International Conference on Information and Computer Technologies (ICICT)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"Fast hardware assisted online learning using unsupervised deep learning structure for anomaly detection\",\"authors\":\"Khaled Alrawashdeh, C. Purdy\",\"doi\":\"10.1109/INFOCT.2018.8356855\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Real-time scenarios of deep learning algorithms are challenged by two less frequently addressed issues. The first is data inefficiency i.e., the model requires several epochs of trial and error to converge which makes it impractical to be applied to real-time applications. The Second is the high precision computation load of the deep learning algorithms to achieve high accuracy during training and inference. In this paper, we address these two issues and apply our model to the task of online anomaly detection using FPGA. To address the first issue, we propose a compressed training model for the contrastive divergence algorithm (CD) in the Deep Belief Network (DBN). The goal is to dynamically adjust the training vector according to the feedback from the free energy and the reconstruction error, which allows for better generalization. To address the second issue, we propose a Hybrid-Stochastic-Dynamic-Fixed-Point (HSDFP) method, which provides training environment with high reduction in calculation, area, and power in FPGA. Our framework enables the DBN structure to take actions and detect attacks online. Thus, the network can collect efficient number of training samples and avoid overfitting. We show that (1) our proposed method converges faster than the state-of-the-art deep learning methods, (2) FPGA implementation achieves accelerated inference speed of 0.008ms and a high power efficiency of 37 G-ops/s/W compared to CPU, GPU, and 16-bit fixed-point arithmetic (3) FPGA also achieves minimal degradation in accuracy of 95%, 95.4%, and 97.9% on the benchmark datasets: MNIST, NSL-KDD, and Kyoto datasets.\",\"PeriodicalId\":376443,\"journal\":{\"name\":\"2018 International Conference on Information and Computer Technologies (ICICT)\",\"volume\":\"41 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 International Conference on Information and Computer Technologies (ICICT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/INFOCT.2018.8356855\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 International Conference on Information and Computer Technologies (ICICT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INFOCT.2018.8356855","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10

摘要

深度学习算法的实时场景受到两个不太经常解决的问题的挑战。首先是数据效率低下,即该模型需要多次尝试和错误才能收敛,这使得它无法应用于实时应用。二是深度学习算法的高精度计算量,在训练和推理过程中达到较高的精度。在本文中,我们解决了这两个问题,并将我们的模型应用于FPGA在线异常检测任务。为了解决第一个问题,我们提出了深度信念网络(DBN)中对比发散算法(CD)的压缩训练模型。目标是根据自由能反馈和重构误差动态调整训练向量,从而实现更好的泛化。为了解决第二个问题,我们提出了一种混合-随机-动态定点(HSDFP)方法,该方法为FPGA提供了计算,面积和功耗大幅减少的训练环境。我们的框架使DBN结构能够在线采取行动并检测攻击。因此,网络可以有效地收集训练样本数量,避免过拟合。我们表明:(1)我们提出的方法比最先进的深度学习方法收敛得更快;(2)与CPU、GPU和16位点算法相比,FPGA实现实现了0.008ms的加速推理速度和37 G-ops/s/W的高功率效率;(3)FPGA在基准数据集:MNIST、NSL-KDD和京都数据集上也实现了95%、95.4%和97.9%的最小精度下降。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Fast hardware assisted online learning using unsupervised deep learning structure for anomaly detection
Real-time scenarios of deep learning algorithms are challenged by two less frequently addressed issues. The first is data inefficiency i.e., the model requires several epochs of trial and error to converge which makes it impractical to be applied to real-time applications. The Second is the high precision computation load of the deep learning algorithms to achieve high accuracy during training and inference. In this paper, we address these two issues and apply our model to the task of online anomaly detection using FPGA. To address the first issue, we propose a compressed training model for the contrastive divergence algorithm (CD) in the Deep Belief Network (DBN). The goal is to dynamically adjust the training vector according to the feedback from the free energy and the reconstruction error, which allows for better generalization. To address the second issue, we propose a Hybrid-Stochastic-Dynamic-Fixed-Point (HSDFP) method, which provides training environment with high reduction in calculation, area, and power in FPGA. Our framework enables the DBN structure to take actions and detect attacks online. Thus, the network can collect efficient number of training samples and avoid overfitting. We show that (1) our proposed method converges faster than the state-of-the-art deep learning methods, (2) FPGA implementation achieves accelerated inference speed of 0.008ms and a high power efficiency of 37 G-ops/s/W compared to CPU, GPU, and 16-bit fixed-point arithmetic (3) FPGA also achieves minimal degradation in accuracy of 95%, 95.4%, and 97.9% on the benchmark datasets: MNIST, NSL-KDD, and Kyoto datasets.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信