Haoyu Wang, Xihe Qiu, Bin Li, Xiaoyu Tan, Jingjing Huang
{"title":"Multimodal heterogeneous graph fusion for automated obstructive sleep apnea-hypopnea syndrome diagnosis","authors":"Haoyu Wang, Xihe Qiu, Bin Li, Xiaoyu Tan, Jingjing Huang","doi":"10.1007/s40747-024-01648-0","DOIUrl":null,"url":null,"abstract":"<p>Polysomnography is the diagnostic gold standard for obstructive sleep apnea-hypopnea syndrome (OSAHS), requiring medical professionals to analyze apnea-hypopnea events from multidimensional data throughout the sleep cycle. This complex process is susceptible to variability based on the clinician’s experience, leading to potential inaccuracies. Existing automatic diagnosis methods often overlook multimodal physiological signals and medical prior knowledge, leading to limited diagnostic capabilities. This study presents a novel <b>hetero</b>geneous <b>g</b>raph <b>c</b>onvolutional <b>f</b>usion <b>net</b>work (<b>HeteroGCFNet</b>) leveraging multimodal physiological signals and domain knowledge for automated OSAHS diagnosis. This framework constructs two types of graph representations: physical space graphs, which map the spatial layout of sensors on the human body, and process knowledge graphs which detail the physiological relationships among breathing patterns, oxygen saturation, and vital signals. The framework leverages heterogeneous graph convolutional neural networks to extract both localized and global features from these graphs. Additionally, a multi-head fusion module combines these features into a unified representation for effective classification, enhancing focus on relevant signal characteristics and cross-modal interactions. This study evaluated the proposed framework on a large-scale OSAHS dataset, combined from publicly available sources and data provided by a collaborative university hospital. It demonstrated superior diagnostic performance compared to conventional machine learning models and existing deep learning approaches, effectively integrating domain knowledge with data-driven learning to produce explainable representations and robust generalization capabilities, which can potentially be utilized for clinical use. Code is available at https://github.com/AmbitYuki/HeteroGCFNet.</p>","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"17 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2024-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex & Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s40747-024-01648-0","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Polysomnography is the diagnostic gold standard for obstructive sleep apnea-hypopnea syndrome (OSAHS), requiring medical professionals to analyze apnea-hypopnea events from multidimensional data throughout the sleep cycle. This complex process is susceptible to variability based on the clinician’s experience, leading to potential inaccuracies. Existing automatic diagnosis methods often overlook multimodal physiological signals and medical prior knowledge, leading to limited diagnostic capabilities. This study presents a novel heterogeneous graph convolutional fusion network (HeteroGCFNet) leveraging multimodal physiological signals and domain knowledge for automated OSAHS diagnosis. This framework constructs two types of graph representations: physical space graphs, which map the spatial layout of sensors on the human body, and process knowledge graphs which detail the physiological relationships among breathing patterns, oxygen saturation, and vital signals. The framework leverages heterogeneous graph convolutional neural networks to extract both localized and global features from these graphs. Additionally, a multi-head fusion module combines these features into a unified representation for effective classification, enhancing focus on relevant signal characteristics and cross-modal interactions. This study evaluated the proposed framework on a large-scale OSAHS dataset, combined from publicly available sources and data provided by a collaborative university hospital. It demonstrated superior diagnostic performance compared to conventional machine learning models and existing deep learning approaches, effectively integrating domain knowledge with data-driven learning to produce explainable representations and robust generalization capabilities, which can potentially be utilized for clinical use. Code is available at https://github.com/AmbitYuki/HeteroGCFNet.
期刊介绍:
Complex & Intelligent Systems aims to provide a forum for presenting and discussing novel approaches, tools and techniques meant for attaining a cross-fertilization between the broad fields of complex systems, computational simulation, and intelligent analytics and visualization. The transdisciplinary research that the journal focuses on will expand the boundaries of our understanding by investigating the principles and processes that underlie many of the most profound problems facing society today.