Ruiyu Zhao , Ian Daly , Yixin Chen , Weijie Wu , Lifei Liu , Xingyu Wang , Andrzej Cichocki , Jing Jin
{"title":"MSAttNet: Multi-scale attention convolutional neural network for motor imagery classification","authors":"Ruiyu Zhao , Ian Daly , Yixin Chen , Weijie Wu , Lifei Liu , Xingyu Wang , Andrzej Cichocki , Jing Jin","doi":"10.1016/j.jneumeth.2025.110578","DOIUrl":null,"url":null,"abstract":"<div><h3>Background:</h3><div>Convolutional neural networks (CNNs) are widely employed in motor imagery (MI) classification. However, due to cumbersome data collection experiments, and limited, noisy, and non-stationary EEG signals, small MI datasets present considerable challenges to the design of these decoding algorithms.</div></div><div><h3>New method:</h3><div>To capture more feature information from inadequately sized data, we propose a new method, a multi-scale attention convolutional neural network (MSAttNet). Our method includes three main components–a multi-band segmentation module, an attention spatial convolution module, and a multi-scale temporal convolution module. First, the multi-band segmentation module adopts a filter bank with overlapping frequency bands to enhance features in the frequency domain. Then, the attention spatial convolution module is used to adaptively adjust different convolutional kernel parameters according to the input through the attention mechanism to capture the features of different datasets. The outputs of the attention spatial convolution module are grouped to perform multi-scale temporal convolution. Finally, the output of the multi-scale temporal convolution module uses the bilinear pooling layer to extract temporal features and perform noise elimination. The extracted features are then classified.</div></div><div><h3>Results:</h3><div>We use four datasets, including <em>BCI Competition IV Dataset IIa</em>, <em>BCI Competition IV Dataset IIb</em>, the <em>OpenBMI</em> dataset and the <em>ECUST-MI</em> dataset, to test our proposed method. MSAttNet achieves accuracies of 78.20%, 84.52%, 75.94% and 78.60% in cross-session experiments, respectively.</div></div><div><h3>Comparison with existing methods</h3><div>: Compared with state-of-the-art algorithms, MSAttNet enhances the decoding performance of MI tasks.</div></div><div><h3>Conclusion:</h3><div>MSAttNet effectively addresses the challenges of MI-EEG datasets, improving decoding performance by robust feature extraction.</div></div>","PeriodicalId":16415,"journal":{"name":"Journal of Neuroscience Methods","volume":"424 ","pages":"Article 110578"},"PeriodicalIF":2.3000,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Neuroscience Methods","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0165027025002225","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"BIOCHEMICAL RESEARCH METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Background:
Convolutional neural networks (CNNs) are widely employed in motor imagery (MI) classification. However, due to cumbersome data collection experiments, and limited, noisy, and non-stationary EEG signals, small MI datasets present considerable challenges to the design of these decoding algorithms.
New method:
To capture more feature information from inadequately sized data, we propose a new method, a multi-scale attention convolutional neural network (MSAttNet). Our method includes three main components–a multi-band segmentation module, an attention spatial convolution module, and a multi-scale temporal convolution module. First, the multi-band segmentation module adopts a filter bank with overlapping frequency bands to enhance features in the frequency domain. Then, the attention spatial convolution module is used to adaptively adjust different convolutional kernel parameters according to the input through the attention mechanism to capture the features of different datasets. The outputs of the attention spatial convolution module are grouped to perform multi-scale temporal convolution. Finally, the output of the multi-scale temporal convolution module uses the bilinear pooling layer to extract temporal features and perform noise elimination. The extracted features are then classified.
Results:
We use four datasets, including BCI Competition IV Dataset IIa, BCI Competition IV Dataset IIb, the OpenBMI dataset and the ECUST-MI dataset, to test our proposed method. MSAttNet achieves accuracies of 78.20%, 84.52%, 75.94% and 78.60% in cross-session experiments, respectively.
Comparison with existing methods
: Compared with state-of-the-art algorithms, MSAttNet enhances the decoding performance of MI tasks.
Conclusion:
MSAttNet effectively addresses the challenges of MI-EEG datasets, improving decoding performance by robust feature extraction.
背景:卷积神经网络(cnn)被广泛应用于运动意象(MI)分类。然而,由于繁琐的数据收集实验和有限的、有噪声的、非平稳的脑电图信号,小MI数据集对这些解码算法的设计提出了相当大的挑战。新方法:为了从尺寸不足的数据中捕获更多的特征信息,我们提出了一种新的方法,即多尺度注意力卷积神经网络(MSAttNet)。该方法包括三个主要部分:多波段分割模块、注意空间卷积模块和多尺度时间卷积模块。首先,多频带分割模块采用频带重叠的滤波器组增强频域特征;然后,利用注意空间卷积模块,通过注意机制,根据输入自适应调整不同的卷积核参数,捕捉不同数据集的特征;对注意空间卷积模块的输出进行分组,进行多尺度时间卷积。最后,多尺度时间卷积模块的输出使用双线性池化层提取时间特征并进行消噪。然后对提取的特征进行分类。结果:我们使用了四个数据集,包括BCI Competition IV Dataset IIa、BCI Competition IV Dataset IIb、OpenBMI数据集和ECUST-MI数据集来测试我们提出的方法。在跨会话实验中,MSAttNet的准确率分别为78.20%、84.52%、75.94%和78.60%。与现有方法的比较:与现有算法相比,MSAttNet提高了MI任务的解码性能。结论:MSAttNet有效地解决了MI-EEG数据集的挑战,通过鲁棒性特征提取提高了解码性能。
期刊介绍:
The Journal of Neuroscience Methods publishes papers that describe new methods that are specifically for neuroscience research conducted in invertebrates, vertebrates or in man. Major methodological improvements or important refinements of established neuroscience methods are also considered for publication. The Journal''s Scope includes all aspects of contemporary neuroscience research, including anatomical, behavioural, biochemical, cellular, computational, molecular, invasive and non-invasive imaging, optogenetic, and physiological research investigations.