Shuaicong Hu , Yanan Wang , Jian Liu , Cuiwei Yang
{"title":"XSleepFusion: A dual-stage information bottleneck fusion framework for interpretable multimodal sleep analysis","authors":"Shuaicong Hu , Yanan Wang , Jian Liu , Cuiwei Yang","doi":"10.1016/j.inffus.2025.103275","DOIUrl":null,"url":null,"abstract":"<div><div>Sleep disorders affect hundreds of millions globally, with accurate assessment of sleep apnea (SA) and sleep staging (SS) essential for clinical diagnosis and early intervention. Manual analysis by sleep experts is time-consuming and subject to inter-rater variability. Deep learning (DL) approaches offer automation potential but face fundamental challenges in multi-modal physiological signal integration and interpretability. This paper presents XSleepFusion, a cross-modal fusion framework based on information bottleneck (IB) theory for automated sleep analysis. The framework introduces a dual-stage IB mechanism that systematically processes physiological signals: first eliminating intra-modal redundancy, then optimizing cross-modal feature fusion. An evolutionary attention Transformer network (EAT-Net) backbone extracts temporal features at multiple scales, providing interpretable attention patterns. Experimental validation on eight clinical datasets comprising over 15,000 sleep recordings demonstrates the framework’s effectiveness in polysomnogram (PSG)-based SA detection, electrocariogram (ECG)-based SA detection, and SS. The architecture achieves superior generalization across varying signal qualities and modal combinations, while the dual-stage design enables flexible integration of diverse physiological signals. Through interpretable feature representations and robust cross-modal fusion capabilities, XSleepFusion establishes a reliable and adaptable foundation for clinical sleep monitoring.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"123 ","pages":"Article 103275"},"PeriodicalIF":14.7000,"publicationDate":"2025-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525003483","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Sleep disorders affect hundreds of millions globally, with accurate assessment of sleep apnea (SA) and sleep staging (SS) essential for clinical diagnosis and early intervention. Manual analysis by sleep experts is time-consuming and subject to inter-rater variability. Deep learning (DL) approaches offer automation potential but face fundamental challenges in multi-modal physiological signal integration and interpretability. This paper presents XSleepFusion, a cross-modal fusion framework based on information bottleneck (IB) theory for automated sleep analysis. The framework introduces a dual-stage IB mechanism that systematically processes physiological signals: first eliminating intra-modal redundancy, then optimizing cross-modal feature fusion. An evolutionary attention Transformer network (EAT-Net) backbone extracts temporal features at multiple scales, providing interpretable attention patterns. Experimental validation on eight clinical datasets comprising over 15,000 sleep recordings demonstrates the framework’s effectiveness in polysomnogram (PSG)-based SA detection, electrocariogram (ECG)-based SA detection, and SS. The architecture achieves superior generalization across varying signal qualities and modal combinations, while the dual-stage design enables flexible integration of diverse physiological signals. Through interpretable feature representations and robust cross-modal fusion capabilities, XSleepFusion establishes a reliable and adaptable foundation for clinical sleep monitoring.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.