Benjamin Fox, Joy Jiang, Sajila Wickramaratne, Patricia Kovatch, Mayte Suarez-Farinas, Neomi A Shah, Ankit Parekh, Girish N Nadkarni
{"title":"A foundational transformer leveraging full night, multichannel sleep study data accurately classifies sleep stages.","authors":"Benjamin Fox, Joy Jiang, Sajila Wickramaratne, Patricia Kovatch, Mayte Suarez-Farinas, Neomi A Shah, Ankit Parekh, Girish N Nadkarni","doi":"10.1093/sleep/zsaf061","DOIUrl":null,"url":null,"abstract":"<p><strong>Study objectives: </strong>To evaluate whether a foundational transformer using 8-hour, multichannel polysomnogram (PSG) data can effectively encode signals and classify sleep stages with state-of-the-art performance.</p><p><strong>Methods: </strong>The Sleep Heart Health Study, Wisconsin Sleep Cohort, and Osteoporotic Fractures in Men (MrOS) Study Visit 1 were used for training, and the Multi-Ethnic Study of Atherosclerosis (MESA), Apnea Positive Pressure Long-term Efficacy Study (APPLES), and MrOS visit 2 served as independent test sets. We developed PFTSleep, a self-supervised foundational transformer that encodes full night sleep studies with brain, movement, cardiac, oxygen, and respiratory channels. These representations were used to train another model to classify sleep stages. We compared our results to existing methods, examined differences in performance by varying channel input data and training dataset size, and investigated an AI explainability tool to analyze decision processes.</p><p><strong>Results: </strong>PFTSleep was trained with 13,888 sleep studies and tested on 4,169 independent studies. Cohen's Kappa scores were 0.81 for our held-out set, 0.59 for APPLES, 0.60 for MESA, and 0.75 for MrOS Visit 2. Performance increases to 0.76 on a held-out MESA set when MESA is included in the training of the classifier head but not the transformer. Compared to other state-of-the-art AI models, our model shows high performance across diverse datasets while only using task agnostic PSG representations from a foundational transformer as input for sleep stage classification.</p><p><strong>Conclusions: </strong>Full night, multichannel PSG representations from a foundational transformer enable accurate sleep stage classification comparable to state-of-the-art AI methods across diverse datasets.</p>","PeriodicalId":22018,"journal":{"name":"Sleep","volume":" ","pages":""},"PeriodicalIF":5.6000,"publicationDate":"2025-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Sleep","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1093/sleep/zsaf061","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0
Abstract
Study objectives: To evaluate whether a foundational transformer using 8-hour, multichannel polysomnogram (PSG) data can effectively encode signals and classify sleep stages with state-of-the-art performance.
Methods: The Sleep Heart Health Study, Wisconsin Sleep Cohort, and Osteoporotic Fractures in Men (MrOS) Study Visit 1 were used for training, and the Multi-Ethnic Study of Atherosclerosis (MESA), Apnea Positive Pressure Long-term Efficacy Study (APPLES), and MrOS visit 2 served as independent test sets. We developed PFTSleep, a self-supervised foundational transformer that encodes full night sleep studies with brain, movement, cardiac, oxygen, and respiratory channels. These representations were used to train another model to classify sleep stages. We compared our results to existing methods, examined differences in performance by varying channel input data and training dataset size, and investigated an AI explainability tool to analyze decision processes.
Results: PFTSleep was trained with 13,888 sleep studies and tested on 4,169 independent studies. Cohen's Kappa scores were 0.81 for our held-out set, 0.59 for APPLES, 0.60 for MESA, and 0.75 for MrOS Visit 2. Performance increases to 0.76 on a held-out MESA set when MESA is included in the training of the classifier head but not the transformer. Compared to other state-of-the-art AI models, our model shows high performance across diverse datasets while only using task agnostic PSG representations from a foundational transformer as input for sleep stage classification.
Conclusions: Full night, multichannel PSG representations from a foundational transformer enable accurate sleep stage classification comparable to state-of-the-art AI methods across diverse datasets.
期刊介绍:
SLEEP® publishes findings from studies conducted at any level of analysis, including:
Genes
Molecules
Cells
Physiology
Neural systems and circuits
Behavior and cognition
Self-report
SLEEP® publishes articles that use a wide variety of scientific approaches and address a broad range of topics. These may include, but are not limited to:
Basic and neuroscience studies of sleep and circadian mechanisms
In vitro and animal models of sleep, circadian rhythms, and human disorders
Pre-clinical human investigations, including the measurement and manipulation of sleep and circadian rhythms
Studies in clinical or population samples. These may address factors influencing sleep and circadian rhythms (e.g., development and aging, and social and environmental influences) and relationships between sleep, circadian rhythms, health, and disease
Clinical trials, epidemiology studies, implementation, and dissemination research.