A foundational transformer leveraging full night, multichannel sleep study data accurately classifies sleep stages

Benjamin Fox, Joy Jiang, Sajila Wickramaratne, Patricia Kovatch, Mayte Suarez-Farinas, Neomi A Shah, Ankit Parekh, Girish N Nadkarni
{"title":"A foundational transformer leveraging full night, multichannel sleep study data accurately classifies sleep stages","authors":"Benjamin Fox, Joy Jiang, Sajila Wickramaratne, Patricia Kovatch, Mayte Suarez-Farinas, Neomi A Shah, Ankit Parekh, Girish N Nadkarni","doi":"10.1101/2024.08.02.24311417","DOIUrl":null,"url":null,"abstract":"Study Objectives. To investigate whether a foundational transformer model using 8-hour, multichannel data from polysomnograms can outperform existing artificial intelligence (AI) methods for sleep stage classification. Methods. We utilized the Sleep Heart Health Study (SHHS) visits 1 and 2 for training and validation and the Multi-Ethnic Study of Atherosclerosis (MESA) for testing of our model. We trained a self-supervised foundational transformer (called PFTSleep) that encodes 8-hour long sleep studies at 125 Hz with 7 signals including brain, movement, cardiac, oxygen, and respiratory channels. These encodings are used as input for training of an additional model to classify sleep stages, without adjusting the weights of the foundational transformer. We compared our results to existing AI methods that did not utilize 8-hour data or the full set of signals but did report evaluation metrics for the SHHS dataset. Results. We trained and validated a model with 8,444 sleep studies with 7 signals including brain, movement, cardiac, oxygen, and respiratory channels and tested on an additional 2,055 studies. In total, we trained and tested 587,944 hours of sleep study signal data. Area under the precision recall curve (AUPRC) scores were 0.82, 0.40, 0.53, 0.75, and 0.82 and area under the receiving operating characteristics curve (AUROC) scores were 0.99, 0.95, 0.96, 0.98, and 0.99 for wake, N1, N2, N3, and REM, respectively, on the SHHS validation set. For MESA, the AUPRC scores were 0.56, 0.16, 0.40, 0.45, and 0.65 and AUROC scores were 0.94, 0.77, 0.87, 0.91, and 0.96, respectively. Our model was compared to the longest context window state-of-the-art model and showed increases in macro evaluation scores, notably sensitivity (3.65% increase) and multi-class REM (3.39% increase) and wake (0.97% increase) F1 scores. Conclusions. Utilizing full night, multi-channel PSG data encodings derived from a foundational transformer improve sleep stage classification over existing methods.","PeriodicalId":501276,"journal":{"name":"medRxiv - Public and Global Health","volume":"8 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"medRxiv - Public and Global Health","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2024.08.02.24311417","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Study Objectives. To investigate whether a foundational transformer model using 8-hour, multichannel data from polysomnograms can outperform existing artificial intelligence (AI) methods for sleep stage classification. Methods. We utilized the Sleep Heart Health Study (SHHS) visits 1 and 2 for training and validation and the Multi-Ethnic Study of Atherosclerosis (MESA) for testing of our model. We trained a self-supervised foundational transformer (called PFTSleep) that encodes 8-hour long sleep studies at 125 Hz with 7 signals including brain, movement, cardiac, oxygen, and respiratory channels. These encodings are used as input for training of an additional model to classify sleep stages, without adjusting the weights of the foundational transformer. We compared our results to existing AI methods that did not utilize 8-hour data or the full set of signals but did report evaluation metrics for the SHHS dataset. Results. We trained and validated a model with 8,444 sleep studies with 7 signals including brain, movement, cardiac, oxygen, and respiratory channels and tested on an additional 2,055 studies. In total, we trained and tested 587,944 hours of sleep study signal data. Area under the precision recall curve (AUPRC) scores were 0.82, 0.40, 0.53, 0.75, and 0.82 and area under the receiving operating characteristics curve (AUROC) scores were 0.99, 0.95, 0.96, 0.98, and 0.99 for wake, N1, N2, N3, and REM, respectively, on the SHHS validation set. For MESA, the AUPRC scores were 0.56, 0.16, 0.40, 0.45, and 0.65 and AUROC scores were 0.94, 0.77, 0.87, 0.91, and 0.96, respectively. Our model was compared to the longest context window state-of-the-art model and showed increases in macro evaluation scores, notably sensitivity (3.65% increase) and multi-class REM (3.39% increase) and wake (0.97% increase) F1 scores. Conclusions. Utilizing full night, multi-channel PSG data encodings derived from a foundational transformer improve sleep stage classification over existing methods.
利用整夜多通道睡眠研究数据准确划分睡眠阶段的基础转换器
研究目的。研究使用多导睡眠图 8 小时多通道数据的基础转换器模型在睡眠阶段分类方面是否优于现有的人工智能(AI)方法。研究方法我们利用睡眠心脏健康研究(SHHS)第 1 次和第 2 次数据对模型进行训练和验证,并利用多种族动脉粥样硬化研究(MESA)数据对模型进行测试。我们训练了一个自我监督的基础转换器(称为 PFTSleep),该转换器以 125 Hz 的频率对 8 小时长的睡眠研究进行编码,其中包括大脑、运动、心脏、氧气和呼吸通道等 7 种信号。这些编码被用作训练附加模型的输入,以对睡眠阶段进行分类,而无需调整基础变换器的权重。我们将我们的结果与现有的人工智能方法进行了比较,这些方法没有使用 8 小时数据或全套信号,但报告了 SHHS 数据集的评估指标。结果我们利用 8444 项睡眠研究和 7 种信号(包括大脑、运动、心脏、氧气和呼吸通道)训练和验证了一个模型,并对另外 2055 项研究进行了测试。我们总共训练和测试了 587,944 小时的睡眠研究信号数据。在 SHHS 验证集上,唤醒、N1、N2、N3 和快速动眼期的精确召回曲线下面积 (AUPRC) 得分分别为 0.82、0.40、0.53、0.75 和 0.82,接收操作特征曲线下面积 (AUROC) 得分分别为 0.99、0.95、0.96、0.98 和 0.99。对于 MESA,AUPRC 分数分别为 0.56、0.16、0.40、0.45 和 0.65,AUROC 分数分别为 0.94、0.77、0.87、0.91 和 0.96。我们的模型与最长上下文窗口的最先进模型进行了比较,结果显示宏观评估分数有所提高,尤其是灵敏度(提高 3.65%)、多类 REM(提高 3.39%)和唤醒(提高 0.97%)F1 分数。结论与现有方法相比,利用基础变换器得出的整夜多通道 PSG 数据编码可改善睡眠阶段分类。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信