{"title":"基于脑电图的混合多模态人机界面的深度学习综述:应用、进展和挑战。","authors":"Hyung-Tak Lee, Miseon Shim, Xianghong Liu, Hye-Ran Cheon, Sang-Gyu Kim, Chang-Hee Han, Han-Jeong Hwang","doi":"10.1007/s13534-025-00469-5","DOIUrl":null,"url":null,"abstract":"<p><p>Human-computer interaction (HCI) focuses on designing efficient and intuitive interactions between humans and computer systems. Recent advancements have utilized multimodal approaches, such as electroencephalography (EEG)-based systems combined with other biosignals, along with deep learning to enhance performance and reliability. However, no systematic review has consolidated findings on EEG-based multimodal HCI systems. This review examined 124 studies published from 2016 to 2024, retrieved from the Web of Science database, focusing on hybrid EEG-based multimodal HCI systems employing deep learning. The keywords used for evaluation were as follows: 'Deep Learning' AND 'EEG' AND ('fNIRS' OR 'NIRS' OR 'MEG' OR 'fMRI' OR 'EOG' OR 'EMG' OR 'ECG' OR 'PPG' OR 'GSR'). Main topics explored are: (1) types of biosignals used with EEG, (2) neural network architectures, (3) fusion strategies, (4) system performance, and (5) target applications. Frequently paired signals, such as EOG, EMG, and fNIRS, effectively complement EEG by addressing its limitations. Convolutional neural networks are extensively employed for spatio-temporal-spectral feature extraction, with early and intermediate fusion strategies being the most commonly used. Applications, such as sleep stage classification, emotion recognition, and mental state decoding, have shown notable performance improvements. Despite these advancements, challenges remain, including the lack of real-time online systems, difficulties in signal synchronization, limited data availability, and insufficient explainable AI (XAI) methods to interpret signal interactions. Emerging solutions, such as portable systems, lightweight deep learning models, and data augmentation techniques, offer promising pathways to address these issues. This review highlights the potential of EEG-based multimodal HCI systems and emphasizes the need for advancements in real-time interaction, fusion algorithms, and XAI to enhance their adaptability, interpretability, and reliability.</p>","PeriodicalId":46898,"journal":{"name":"Biomedical Engineering Letters","volume":"15 4","pages":"587-618"},"PeriodicalIF":3.2000,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229323/pdf/","citationCount":"0","resultStr":"{\"title\":\"A review of hybrid EEG-based multimodal human-computer interfaces using deep learning: applications, advances, and challenges.\",\"authors\":\"Hyung-Tak Lee, Miseon Shim, Xianghong Liu, Hye-Ran Cheon, Sang-Gyu Kim, Chang-Hee Han, Han-Jeong Hwang\",\"doi\":\"10.1007/s13534-025-00469-5\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Human-computer interaction (HCI) focuses on designing efficient and intuitive interactions between humans and computer systems. Recent advancements have utilized multimodal approaches, such as electroencephalography (EEG)-based systems combined with other biosignals, along with deep learning to enhance performance and reliability. However, no systematic review has consolidated findings on EEG-based multimodal HCI systems. This review examined 124 studies published from 2016 to 2024, retrieved from the Web of Science database, focusing on hybrid EEG-based multimodal HCI systems employing deep learning. The keywords used for evaluation were as follows: 'Deep Learning' AND 'EEG' AND ('fNIRS' OR 'NIRS' OR 'MEG' OR 'fMRI' OR 'EOG' OR 'EMG' OR 'ECG' OR 'PPG' OR 'GSR'). Main topics explored are: (1) types of biosignals used with EEG, (2) neural network architectures, (3) fusion strategies, (4) system performance, and (5) target applications. Frequently paired signals, such as EOG, EMG, and fNIRS, effectively complement EEG by addressing its limitations. Convolutional neural networks are extensively employed for spatio-temporal-spectral feature extraction, with early and intermediate fusion strategies being the most commonly used. Applications, such as sleep stage classification, emotion recognition, and mental state decoding, have shown notable performance improvements. Despite these advancements, challenges remain, including the lack of real-time online systems, difficulties in signal synchronization, limited data availability, and insufficient explainable AI (XAI) methods to interpret signal interactions. Emerging solutions, such as portable systems, lightweight deep learning models, and data augmentation techniques, offer promising pathways to address these issues. This review highlights the potential of EEG-based multimodal HCI systems and emphasizes the need for advancements in real-time interaction, fusion algorithms, and XAI to enhance their adaptability, interpretability, and reliability.</p>\",\"PeriodicalId\":46898,\"journal\":{\"name\":\"Biomedical Engineering Letters\",\"volume\":\"15 4\",\"pages\":\"587-618\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-03-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12229323/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biomedical Engineering Letters\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1007/s13534-025-00469-5\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/7/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, BIOMEDICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical Engineering Letters","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1007/s13534-025-00469-5","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/7/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0
摘要
人机交互(HCI)侧重于设计人与计算机系统之间有效和直观的交互。最近的进展是利用多模式方法,例如基于脑电图(EEG)的系统与其他生物信号相结合,以及深度学习来提高性能和可靠性。然而,目前还没有系统性的综述对基于脑电图的多模态HCI系统的研究结果进行整合。本综述从Web of Science数据库中检索了2016年至2024年发表的124项研究,重点关注采用深度学习的基于脑电图的混合多模态HCI系统。用于评估的关键词如下:“深度学习”和“脑电图”和(“fNIRS”或“NIRS”或“MEG”或“fMRI”或“EOG”或“EMG”或“ECG”或“PPG”或“GSR”)。主要探讨的主题有:(1)脑电图使用的生物信号类型,(2)神经网络架构,(3)融合策略,(4)系统性能,以及(5)目标应用。频繁配对的信号,如EOG、EMG和fNIRS,通过解决EEG的局限性,有效地补充了EEG。卷积神经网络广泛用于时空光谱特征提取,其中早期和中期融合策略是最常用的。睡眠阶段分类、情绪识别和精神状态解码等应用已经显示出显著的性能改善。尽管取得了这些进步,但挑战依然存在,包括缺乏实时在线系统、信号同步困难、数据可用性有限以及解释信号相互作用的可解释人工智能(XAI)方法不足。新兴的解决方案,如便携式系统、轻量级深度学习模型和数据增强技术,为解决这些问题提供了有希望的途径。这篇综述强调了基于脑电图的多模态HCI系统的潜力,并强调需要在实时交互、融合算法和XAI方面取得进展,以增强其适应性、可解释性和可靠性。
A review of hybrid EEG-based multimodal human-computer interfaces using deep learning: applications, advances, and challenges.
Human-computer interaction (HCI) focuses on designing efficient and intuitive interactions between humans and computer systems. Recent advancements have utilized multimodal approaches, such as electroencephalography (EEG)-based systems combined with other biosignals, along with deep learning to enhance performance and reliability. However, no systematic review has consolidated findings on EEG-based multimodal HCI systems. This review examined 124 studies published from 2016 to 2024, retrieved from the Web of Science database, focusing on hybrid EEG-based multimodal HCI systems employing deep learning. The keywords used for evaluation were as follows: 'Deep Learning' AND 'EEG' AND ('fNIRS' OR 'NIRS' OR 'MEG' OR 'fMRI' OR 'EOG' OR 'EMG' OR 'ECG' OR 'PPG' OR 'GSR'). Main topics explored are: (1) types of biosignals used with EEG, (2) neural network architectures, (3) fusion strategies, (4) system performance, and (5) target applications. Frequently paired signals, such as EOG, EMG, and fNIRS, effectively complement EEG by addressing its limitations. Convolutional neural networks are extensively employed for spatio-temporal-spectral feature extraction, with early and intermediate fusion strategies being the most commonly used. Applications, such as sleep stage classification, emotion recognition, and mental state decoding, have shown notable performance improvements. Despite these advancements, challenges remain, including the lack of real-time online systems, difficulties in signal synchronization, limited data availability, and insufficient explainable AI (XAI) methods to interpret signal interactions. Emerging solutions, such as portable systems, lightweight deep learning models, and data augmentation techniques, offer promising pathways to address these issues. This review highlights the potential of EEG-based multimodal HCI systems and emphasizes the need for advancements in real-time interaction, fusion algorithms, and XAI to enhance their adaptability, interpretability, and reliability.
期刊介绍:
Biomedical Engineering Letters (BMEL) aims to present the innovative experimental science and technological development in the biomedical field as well as clinical application of new development. The article must contain original biomedical engineering content, defined as development, theoretical analysis, and evaluation/validation of a new technique. BMEL publishes the following types of papers: original articles, review articles, editorials, and letters to the editor. All the papers are reviewed in single-blind fashion.