Multiclass Classification of Imagined Speech Vowels and Words of Electroencephalography Signals Using Deep Learning

Nrushingh Charan Mahapatra, Prachet Bhuyan
{"title":"Multiclass Classification of Imagined Speech Vowels and Words of Electroencephalography Signals Using Deep Learning","authors":"Nrushingh Charan Mahapatra, Prachet Bhuyan","doi":"10.1155/2022/1374880","DOIUrl":null,"url":null,"abstract":"The paper’s emphasis is on the imagined speech decoding of electroencephalography (EEG) neural signals of individuals in accordance with the expansion of the brain-computer interface to encompass individuals with speech problems encountering communication challenges. Decoding an individual’s imagined speech from nonstationary and nonlinear EEG neural signals is a complex task. Related research work in the field of imagined speech has revealed that imagined speech decoding performance and accuracy require attention to further improve. The evolution of deep learning technology increases the likelihood of decoding imagined speech from EEG signals with enhanced performance. We proposed a novel supervised deep learning model that combined the temporal convolutional networks and the convolutional neural networks with the intent of retrieving information from the EEG signals. The experiment was carried out using an open-access dataset of fifteen subjects’ imagined speech multichannel signals of vowels and words. The raw multichannel EEG signals of multiple subjects were processed using discrete wavelet transformation technique. The model was trained and evaluated using the preprocessed signals, and the model hyperparameters were adjusted to achieve higher accuracy in the classification of imagined speech. The experiment results demonstrated that the multiclass imagined speech classification of the proposed model exhibited a higher overall accuracy of 0.9649 and a classification error rate of 0.0350. The results of the study indicate that individuals with speech difficulties might well be able to leverage a noninvasive EEG-based imagined speech brain-computer interface system as one of the long-term alternative artificial verbal communication mediums.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"60 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Adv. Hum. Comput. Interact.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1155/2022/1374880","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

The paper’s emphasis is on the imagined speech decoding of electroencephalography (EEG) neural signals of individuals in accordance with the expansion of the brain-computer interface to encompass individuals with speech problems encountering communication challenges. Decoding an individual’s imagined speech from nonstationary and nonlinear EEG neural signals is a complex task. Related research work in the field of imagined speech has revealed that imagined speech decoding performance and accuracy require attention to further improve. The evolution of deep learning technology increases the likelihood of decoding imagined speech from EEG signals with enhanced performance. We proposed a novel supervised deep learning model that combined the temporal convolutional networks and the convolutional neural networks with the intent of retrieving information from the EEG signals. The experiment was carried out using an open-access dataset of fifteen subjects’ imagined speech multichannel signals of vowels and words. The raw multichannel EEG signals of multiple subjects were processed using discrete wavelet transformation technique. The model was trained and evaluated using the preprocessed signals, and the model hyperparameters were adjusted to achieve higher accuracy in the classification of imagined speech. The experiment results demonstrated that the multiclass imagined speech classification of the proposed model exhibited a higher overall accuracy of 0.9649 and a classification error rate of 0.0350. The results of the study indicate that individuals with speech difficulties might well be able to leverage a noninvasive EEG-based imagined speech brain-computer interface system as one of the long-term alternative artificial verbal communication mediums.
基于深度学习的脑电信号想象语音元音和词的多类分类
本文的重点是根据脑机接口的扩展,对个体的脑电图(EEG)神经信号进行想象语音解码,以涵盖遇到沟通困难的言语障碍个体。从非平稳和非线性的脑电图神经信号中解码个体的想象语音是一项复杂的任务。在想象语音领域的相关研究表明,想象语音的解码性能和准确率有待进一步提高。深度学习技术的发展提高了从脑电图信号中解码想象语音的可能性,并提高了性能。我们提出了一种新的监督深度学习模型,该模型结合了时间卷积网络和卷积神经网络,目的是从脑电信号中检索信息。实验使用了一个开放获取的数据集,其中包含15名受试者想象的语音多通道信号,包括元音和单词。采用离散小波变换技术对多受试者的原始多通道脑电信号进行处理。利用预处理后的信号对模型进行训练和评估,并调整模型的超参数,以提高对想象语音的分类精度。实验结果表明,该模型的多类想象语音分类总体准确率为0.9649,分类错误率为0.0350。研究结果表明,有语言障碍的人很可能能够利用基于脑电图的无创想象语音脑机接口系统作为长期替代人工语言交流媒介之一。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信