{"title":"Multilevel Inter-modal and Intra-modal Transformer network with domain adversarial learning for multimodal sleep staging.","authors":"Yang-Yang He, Jian-Wei Liu","doi":"10.1007/s11571-025-10262-w","DOIUrl":null,"url":null,"abstract":"<p><p>Sleep staging identification is a fundamental task for the diagnosis of sleep disorders. With the development of biosensing technology and deep learning technology, it is possible to automatically decode sleep process through electroencephalogram signals. However, most sleep staging methods do not consider multimodal sleep signals such as electroencephalogram and electrooculograms signals simultaneously for sleep staging due to the limitation of performance improvement. To this regard, we design a Multilevel Inter-modal and Intra-modal Transformer network with domain adversarial learning for multimodal sleep staging, we introduce a multilevel Transformer structure to fully capture the temporal dependencies within sleep signals of each modality and the interdependencies among different modalities. Simultaneously, we strive for the multi-scale CNNs to learn time and frequency features separately. Our research promotes the application of Transformer models in the field of sleep staging identification. Moreover, considering individual differences among subjects, models trained on one group's data often perform poorly when applied to another group, known as the domain generalization problem. While domain adaptation methods are commonly used, fine-tuning on the target domain each time is cumbersome and impractical. To effectively address these issues without using target domain information, we introduce domain adversarial learning to help the model learn domain-invariant features for better generalization across domains. We validated the superiority of our model on two commonly used datasets, significantly outperforming other baseline models. Our model efficiently extracts dependencies of intra-modal level and inter-modal level from multimodal sleep data, making it suitable for scenarios requiring high accuracy.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"80"},"PeriodicalIF":3.1000,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12106285/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cognitive Neurodynamics","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1007/s11571-025-10262-w","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/5/26 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
Sleep staging identification is a fundamental task for the diagnosis of sleep disorders. With the development of biosensing technology and deep learning technology, it is possible to automatically decode sleep process through electroencephalogram signals. However, most sleep staging methods do not consider multimodal sleep signals such as electroencephalogram and electrooculograms signals simultaneously for sleep staging due to the limitation of performance improvement. To this regard, we design a Multilevel Inter-modal and Intra-modal Transformer network with domain adversarial learning for multimodal sleep staging, we introduce a multilevel Transformer structure to fully capture the temporal dependencies within sleep signals of each modality and the interdependencies among different modalities. Simultaneously, we strive for the multi-scale CNNs to learn time and frequency features separately. Our research promotes the application of Transformer models in the field of sleep staging identification. Moreover, considering individual differences among subjects, models trained on one group's data often perform poorly when applied to another group, known as the domain generalization problem. While domain adaptation methods are commonly used, fine-tuning on the target domain each time is cumbersome and impractical. To effectively address these issues without using target domain information, we introduce domain adversarial learning to help the model learn domain-invariant features for better generalization across domains. We validated the superiority of our model on two commonly used datasets, significantly outperforming other baseline models. Our model efficiently extracts dependencies of intra-modal level and inter-modal level from multimodal sleep data, making it suitable for scenarios requiring high accuracy.
期刊介绍:
Cognitive Neurodynamics provides a unique forum of communication and cooperation for scientists and engineers working in the field of cognitive neurodynamics, intelligent science and applications, bridging the gap between theory and application, without any preference for pure theoretical, experimental or computational models.
The emphasis is to publish original models of cognitive neurodynamics, novel computational theories and experimental results. In particular, intelligent science inspired by cognitive neuroscience and neurodynamics is also very welcome.
The scope of Cognitive Neurodynamics covers cognitive neuroscience, neural computation based on dynamics, computer science, intelligent science as well as their interdisciplinary applications in the natural and engineering sciences. Papers that are appropriate for non-specialist readers are encouraged.
1. There is no page limit for manuscripts submitted to Cognitive Neurodynamics. Research papers should clearly represent an important advance of especially broad interest to researchers and technologists in neuroscience, biophysics, BCI, neural computer and intelligent robotics.
2. Cognitive Neurodynamics also welcomes brief communications: short papers reporting results that are of genuinely broad interest but that for one reason and another do not make a sufficiently complete story to justify a full article publication. Brief Communications should consist of approximately four manuscript pages.
3. Cognitive Neurodynamics publishes review articles in which a specific field is reviewed through an exhaustive literature survey. There are no restrictions on the number of pages. Review articles are usually invited, but submitted reviews will also be considered.