Analysis of the acceleration of deep learning inference models on a heterogeneous architecture based on OpenVINO

Fatima Zahra Guerrouj, Mohamed Abouzahir, M. Ramzi, E. Abdali
{"title":"Analysis of the acceleration of deep learning inference models on a heterogeneous architecture based on OpenVINO","authors":"Fatima Zahra Guerrouj, Mohamed Abouzahir, M. Ramzi, E. Abdali","doi":"10.1109/ISAECT53699.2021.9668607","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks (CNN) are a very powerful tool for many different applications. This capability is highly demanded in the field of embedded systems for video surveillance, speech recognition, and image analysis. Due to its high computational intensity, the application of CNN is limited to real-time research areas where computational speed is extremely important. Therefore, an appropriate accelerator is required to fulfill the requirements of these limitations. On the one hand, GPUs are widely used to accelerate the CNN under high power dissipation. On the other hand, the trend for FPGA implementation is increasing rapidly due to its low power consumption and facile re-configurability. In this work, we evaluate the inference performance of 10 classification models and 9 object detection models using the OpenVINO toolkit. In addition, we analyzed the implementation of these models on the DE5a-Net DDR4 equipped with an Arria 10 GX FPGA. The results show that the performance of Full-Precision FP32 classification models on a heterogeneous architecture FPGA/CPU is on average 3.6X faster than the CPU.","PeriodicalId":137636,"journal":{"name":"2021 4th International Symposium on Advanced Electrical and Communication Technologies (ISAECT)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 4th International Symposium on Advanced Electrical and Communication Technologies (ISAECT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISAECT53699.2021.9668607","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Convolutional neural networks (CNN) are a very powerful tool for many different applications. This capability is highly demanded in the field of embedded systems for video surveillance, speech recognition, and image analysis. Due to its high computational intensity, the application of CNN is limited to real-time research areas where computational speed is extremely important. Therefore, an appropriate accelerator is required to fulfill the requirements of these limitations. On the one hand, GPUs are widely used to accelerate the CNN under high power dissipation. On the other hand, the trend for FPGA implementation is increasing rapidly due to its low power consumption and facile re-configurability. In this work, we evaluate the inference performance of 10 classification models and 9 object detection models using the OpenVINO toolkit. In addition, we analyzed the implementation of these models on the DE5a-Net DDR4 equipped with an Arria 10 GX FPGA. The results show that the performance of Full-Precision FP32 classification models on a heterogeneous architecture FPGA/CPU is on average 3.6X faster than the CPU.
基于OpenVINO的异构架构下深度学习推理模型加速分析
卷积神经网络(CNN)对于许多不同的应用来说是一个非常强大的工具。这种能力在视频监控、语音识别和图像分析的嵌入式系统领域是非常需要的。由于计算强度高,CNN的应用仅限于对计算速度极其重要的实时研究领域。因此,需要适当的加速器来满足这些限制的要求。一方面,gpu被广泛用于在高功耗下加速CNN。另一方面,由于FPGA的低功耗和易于重新配置,其实现的趋势正在迅速增加。在这项工作中,我们使用OpenVINO工具包评估了10个分类模型和9个目标检测模型的推理性能。此外,我们还分析了这些模型在配备Arria 10 GX FPGA的DE5a-Net DDR4上的实现。结果表明,在异构架构FPGA/CPU上,全精度FP32分类模型的性能平均比CPU快3.6倍。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信