Daniel Flood, Neethu Robinson, Shanker Shreejith
{"title":"FPGA-based Deep-Learning Accelerators for Energy Efficient Motor Imagery EEG classification","authors":"Daniel Flood, Neethu Robinson, Shanker Shreejith","doi":"10.1109/COINS54846.2022.9854985","DOIUrl":null,"url":null,"abstract":"In recent years, Deep Learning has emerged as a powerful framework for analysing and decoding bio-signals like Electroencephalography (EEG) with applications in brain computer interfaces (BCI) and motor control. Deep convolutional neural networks have shown to be highly effective in decoding BCI signals for applications like two-class motor imagery decoding. Their deployment in real-time applications, however, requires highly parallel and capable computing platforms like GPUs to achieve high-speed inference, consuming a large amount of energy. In this paper, we explore a custom deep learning accelerator on an off-the-shelf hybrid FPGA device to achieve similar inference performance at a fraction of the energy consumption. We evaluate different optimisations at bit-level, data-path and training using a state-of-the-art deep convolutional neural network as our baseline model to arrive at our custom precision quantised deep learning model, which is implemented using the FINN compiler from Xilinx. The accelerator, deployed on a Xilinx Zynq Ultrascale+ FPGA, achieves a significant reduction in power consumption (≈ 17×), sub 2 ms decoding latency and a near-identical decoding accuracy (statistically insignificant reduction of 2.5% average) as the reported baseline subject-specific classification accuracy on an N (= 54) subject motor imagery EEG (MI-EEG) dataset compared to the Deep CNN model on GPU, making our approach more appealing for low-power real-time BCI applications. Furthermore, this design approach is transferable to other deep learning models reported in BCI research, paving the way for novel applications of real-time portable BCI systems.","PeriodicalId":187055,"journal":{"name":"2022 IEEE International Conference on Omni-layer Intelligent Systems (COINS)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Omni-layer Intelligent Systems (COINS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/COINS54846.2022.9854985","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

近年来,深度学习已成为分析和解码脑电图(EEG)等生物信号的强大框架,并应用于脑机接口(BCI)和运动控制。深度卷积神经网络已被证明在解码BCI信号方面非常有效,例如解码两类运动图像。然而,在实时应用中部署它们需要像gpu这样的高度并行和强大的计算平台来实现高速推理,消耗大量的能量。我们使用最先进的深度卷积神经网络作为基线模型,在位级、数据路径和训练方面评估不同的优化,以达到我们定制的精确量化深度学习模型,该模型使用Xilinx的FINN编译器实现。该加速器部署在Xilinx Zynq Ultrascale+ FPGA上,与GPU上的Deep CNN模型相比,在N(= 54)个受试者运动图像脑电图(MI-EEG)数据集上实现了功耗显著降低(≈17×),解码延迟低于2 ms,解码精度几乎相同(统计上平均降低2.5%),使我们的方法对低功耗实时脑机接口应用更具吸引力。此外,这种设计方法可转移到BCI研究中报道的其他深度学习模型,为实时便携式BCI系统的新应用铺平了道路。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
FPGA-based Deep-Learning Accelerators for Energy Efficient Motor Imagery EEG classification
In recent years, Deep Learning has emerged as a powerful framework for analysing and decoding bio-signals like Electroencephalography (EEG) with applications in brain computer interfaces (BCI) and motor control. Deep convolutional neural networks have shown to be highly effective in decoding BCI signals for applications like two-class motor imagery decoding. Their deployment in real-time applications, however, requires highly parallel and capable computing platforms like GPUs to achieve high-speed inference, consuming a large amount of energy. In this paper, we explore a custom deep learning accelerator on an off-the-shelf hybrid FPGA device to achieve similar inference performance at a fraction of the energy consumption. We evaluate different optimisations at bit-level, data-path and training using a state-of-the-art deep convolutional neural network as our baseline model to arrive at our custom precision quantised deep learning model, which is implemented using the FINN compiler from Xilinx. The accelerator, deployed on a Xilinx Zynq Ultrascale+ FPGA, achieves a significant reduction in power consumption (≈ 17×), sub 2 ms decoding latency and a near-identical decoding accuracy (statistically insignificant reduction of 2.5% average) as the reported baseline subject-specific classification accuracy on an N (= 54) subject motor imagery EEG (MI-EEG) dataset compared to the Deep CNN model on GPU, making our approach more appealing for low-power real-time BCI applications. Furthermore, this design approach is transferable to other deep learning models reported in BCI research, paving the way for novel applications of real-time portable BCI systems.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信