{"title":"[基于动态多尺度卷积和多头时间注意的运动意象分类]。","authors":"Nan Xiao, Ming'ai Li","doi":"10.7507/1001-5515.202408051","DOIUrl":null,"url":null,"abstract":"<p><p>Convolutional neural networks (CNNs) are renowned for their excellent representation learning capabilities and have become a mainstream model for motor imagery based electroencephalogram (MI-EEG) signal classification. However, MI-EEG exhibits strong inter-individual variability, which may lead to a decline in classification performance. To address this issue, this paper proposes a classification model based on dynamic multi-scale CNN and multi-head temporal attention (DMSCMHTA). The model first applies multi-band filtering to the raw MI-EEG signals and inputs the results into the feature extraction module. Then, it uses a dynamic multi-scale CNN to capture temporal features while adjusting attention weights, followed by spatial convolution to extract spatiotemporal feature sequences. Next, the model further optimizes temporal correlations through time dimensionality reduction and a multi-head attention mechanism to generate more discriminative features. Finally, MI classification is completed under the supervision of cross-entropy loss and center loss. Experiments show that the proposed model achieves average accuracies of 80.32% and 90.81% on BCI Competition IV datasets 2a and 2b, respectively. The results indicate that DMSCMHTA can adaptively extract personalized spatiotemporal features and outperforms current mainstream methods.</p>","PeriodicalId":39324,"journal":{"name":"生物医学工程学杂志","volume":"42 4","pages":"678-685"},"PeriodicalIF":0.0000,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12409511/pdf/","citationCount":"0","resultStr":"{\"title\":\"[Motor imagery classification based on dynamic multi-scale convolution and multi-head temporal attention].\",\"authors\":\"Nan Xiao, Ming'ai Li\",\"doi\":\"10.7507/1001-5515.202408051\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Convolutional neural networks (CNNs) are renowned for their excellent representation learning capabilities and have become a mainstream model for motor imagery based electroencephalogram (MI-EEG) signal classification. However, MI-EEG exhibits strong inter-individual variability, which may lead to a decline in classification performance. To address this issue, this paper proposes a classification model based on dynamic multi-scale CNN and multi-head temporal attention (DMSCMHTA). The model first applies multi-band filtering to the raw MI-EEG signals and inputs the results into the feature extraction module. Then, it uses a dynamic multi-scale CNN to capture temporal features while adjusting attention weights, followed by spatial convolution to extract spatiotemporal feature sequences. Next, the model further optimizes temporal correlations through time dimensionality reduction and a multi-head attention mechanism to generate more discriminative features. Finally, MI classification is completed under the supervision of cross-entropy loss and center loss. Experiments show that the proposed model achieves average accuracies of 80.32% and 90.81% on BCI Competition IV datasets 2a and 2b, respectively. The results indicate that DMSCMHTA can adaptively extract personalized spatiotemporal features and outperforms current mainstream methods.</p>\",\"PeriodicalId\":39324,\"journal\":{\"name\":\"生物医学工程学杂志\",\"volume\":\"42 4\",\"pages\":\"678-685\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-08-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12409511/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"生物医学工程学杂志\",\"FirstCategoryId\":\"1087\",\"ListUrlMain\":\"https://doi.org/10.7507/1001-5515.202408051\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"Medicine\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"生物医学工程学杂志","FirstCategoryId":"1087","ListUrlMain":"https://doi.org/10.7507/1001-5515.202408051","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"Medicine","Score":null,"Total":0}
[Motor imagery classification based on dynamic multi-scale convolution and multi-head temporal attention].
Convolutional neural networks (CNNs) are renowned for their excellent representation learning capabilities and have become a mainstream model for motor imagery based electroencephalogram (MI-EEG) signal classification. However, MI-EEG exhibits strong inter-individual variability, which may lead to a decline in classification performance. To address this issue, this paper proposes a classification model based on dynamic multi-scale CNN and multi-head temporal attention (DMSCMHTA). The model first applies multi-band filtering to the raw MI-EEG signals and inputs the results into the feature extraction module. Then, it uses a dynamic multi-scale CNN to capture temporal features while adjusting attention weights, followed by spatial convolution to extract spatiotemporal feature sequences. Next, the model further optimizes temporal correlations through time dimensionality reduction and a multi-head attention mechanism to generate more discriminative features. Finally, MI classification is completed under the supervision of cross-entropy loss and center loss. Experiments show that the proposed model achieves average accuracies of 80.32% and 90.81% on BCI Competition IV datasets 2a and 2b, respectively. The results indicate that DMSCMHTA can adaptively extract personalized spatiotemporal features and outperforms current mainstream methods.