青光眼视野预测的可解释深度学习:伪影校正增强变压器模型。

IF 2.6 3区 医学 Q2 OPHTHALMOLOGY
Kornchanok Sriwatana, Chanon Puttanawarut, Yanin Suwan, Titipat Achakulvisut
{"title":"青光眼视野预测的可解释深度学习:伪影校正增强变压器模型。","authors":"Kornchanok Sriwatana, Chanon Puttanawarut, Yanin Suwan, Titipat Achakulvisut","doi":"10.1167/tvst.14.1.22","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>The purpose of this study was to develop a deep learning approach that restores artifact-laden optical coherence tomography (OCT) scans and predicts functional loss on the 24-2 Humphrey Visual Field (HVF) test.</p><p><strong>Methods: </strong>This cross-sectional, retrospective study used 1674 visual field (VF)-OCT pairs from 951 eyes for training and 429 pairs from 345 eyes for testing. Peripapillary retinal nerve fiber layer (RNFL) thickness map artifacts were corrected using a generative diffusion model. Three convolutional neural networks and 2 transformer-based models were trained on original and artifact-corrected datasets to estimate 54 sensitivity thresholds of the 24-2 HVF test.</p><p><strong>Results: </strong>Predictive performances were calculated using root mean square error (RMSE) and mean absolute error (MAE), with explainability evaluated through GradCAM, attention maps, and dimensionality reduction techniques. The Distillation with No Labels (DINO) Vision Transformers (ViT) trained on artifact-corrected datasets achieved the highest accuracy (RMSE, 95% confidence interval [CI] = 4.44, 95% CI = 4.07, 4.82 decibel [dB], MAE = 3.46, 95% CI = 3.14, 3.79 dB), and the greatest interpretability, showing improvements of 0.15 dB in global RMSE and MAE (P < 0.05) compared to the performance on original maps. Feature maps and visualization tools indicate that artifacts compromise DINO-ViT's predictive ability but improve with artifact correction.</p><p><strong>Conclusions: </strong>Combining self-supervised ViTs with generative artifact correction enhances the correlation between glaucomatous structures and functions.</p><p><strong>Translational relevance: </strong>Our approach offers a comprehensive tool for glaucoma management, facilitates the exploration of structure-function correlations in research, and underscores the importance of addressing artifacts in the clinical interpretation of OCT.</p>","PeriodicalId":23322,"journal":{"name":"Translational Vision Science & Technology","volume":"14 1","pages":"22"},"PeriodicalIF":2.6000,"publicationDate":"2025-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11758932/pdf/","citationCount":"0","resultStr":"{\"title\":\"Explainable Deep Learning for Glaucomatous Visual Field Prediction: Artifact Correction Enhances Transformer Models.\",\"authors\":\"Kornchanok Sriwatana, Chanon Puttanawarut, Yanin Suwan, Titipat Achakulvisut\",\"doi\":\"10.1167/tvst.14.1.22\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Purpose: </strong>The purpose of this study was to develop a deep learning approach that restores artifact-laden optical coherence tomography (OCT) scans and predicts functional loss on the 24-2 Humphrey Visual Field (HVF) test.</p><p><strong>Methods: </strong>This cross-sectional, retrospective study used 1674 visual field (VF)-OCT pairs from 951 eyes for training and 429 pairs from 345 eyes for testing. Peripapillary retinal nerve fiber layer (RNFL) thickness map artifacts were corrected using a generative diffusion model. Three convolutional neural networks and 2 transformer-based models were trained on original and artifact-corrected datasets to estimate 54 sensitivity thresholds of the 24-2 HVF test.</p><p><strong>Results: </strong>Predictive performances were calculated using root mean square error (RMSE) and mean absolute error (MAE), with explainability evaluated through GradCAM, attention maps, and dimensionality reduction techniques. The Distillation with No Labels (DINO) Vision Transformers (ViT) trained on artifact-corrected datasets achieved the highest accuracy (RMSE, 95% confidence interval [CI] = 4.44, 95% CI = 4.07, 4.82 decibel [dB], MAE = 3.46, 95% CI = 3.14, 3.79 dB), and the greatest interpretability, showing improvements of 0.15 dB in global RMSE and MAE (P < 0.05) compared to the performance on original maps. Feature maps and visualization tools indicate that artifacts compromise DINO-ViT's predictive ability but improve with artifact correction.</p><p><strong>Conclusions: </strong>Combining self-supervised ViTs with generative artifact correction enhances the correlation between glaucomatous structures and functions.</p><p><strong>Translational relevance: </strong>Our approach offers a comprehensive tool for glaucoma management, facilitates the exploration of structure-function correlations in research, and underscores the importance of addressing artifacts in the clinical interpretation of OCT.</p>\",\"PeriodicalId\":23322,\"journal\":{\"name\":\"Translational Vision Science & Technology\",\"volume\":\"14 1\",\"pages\":\"22\"},\"PeriodicalIF\":2.6000,\"publicationDate\":\"2025-01-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11758932/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Translational Vision Science & Technology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1167/tvst.14.1.22\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"OPHTHALMOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Translational Vision Science & Technology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1167/tvst.14.1.22","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

目的:本研究的目的是开发一种深度学习方法,以恢复伪影负载光学相干断层扫描(OCT)扫描,并预测24-2汉弗莱视野(HVF)测试中的功能损失。方法:本研究采用951只眼的1674对视野(VF)-OCT进行训练,345只眼的429对进行测试。采用生成扩散模型对乳头周围视网膜神经纤维层(RNFL)厚度图伪影进行校正。在原始和人工校正数据集上训练3个卷积神经网络和2个基于变压器的模型,以估计24-2 HVF测试的54个灵敏度阈值。结果:使用均方根误差(RMSE)和平均绝对误差(MAE)计算预测性能,并通过GradCAM、注意图和降维技术评估可解释性。在人工校正数据集上训练的无标签蒸馏(DINO)视觉变形器(ViT)获得了最高的准确度(RMSE, 95%置信区间[CI] = 4.44, 95% CI = 4.07, 4.82分贝[dB], MAE = 3.46, 95% CI = 3.14, 3.79 dB)和最大的可解释性,与原始地图的性能相比,总体RMSE和MAE提高了0.15 dB (P < 0.05)。特征图和可视化工具表明,伪影损害了DINO-ViT的预测能力,但通过伪影校正得到了改善。结论:自监督ViTs与生成伪影校正相结合,增强了青光眼结构与功能的相关性。翻译相关性:我们的方法为青光眼治疗提供了一个全面的工具,促进了研究中结构-功能相关性的探索,并强调了在OCT临床解释中解决伪影的重要性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Explainable Deep Learning for Glaucomatous Visual Field Prediction: Artifact Correction Enhances Transformer Models.

Purpose: The purpose of this study was to develop a deep learning approach that restores artifact-laden optical coherence tomography (OCT) scans and predicts functional loss on the 24-2 Humphrey Visual Field (HVF) test.

Methods: This cross-sectional, retrospective study used 1674 visual field (VF)-OCT pairs from 951 eyes for training and 429 pairs from 345 eyes for testing. Peripapillary retinal nerve fiber layer (RNFL) thickness map artifacts were corrected using a generative diffusion model. Three convolutional neural networks and 2 transformer-based models were trained on original and artifact-corrected datasets to estimate 54 sensitivity thresholds of the 24-2 HVF test.

Results: Predictive performances were calculated using root mean square error (RMSE) and mean absolute error (MAE), with explainability evaluated through GradCAM, attention maps, and dimensionality reduction techniques. The Distillation with No Labels (DINO) Vision Transformers (ViT) trained on artifact-corrected datasets achieved the highest accuracy (RMSE, 95% confidence interval [CI] = 4.44, 95% CI = 4.07, 4.82 decibel [dB], MAE = 3.46, 95% CI = 3.14, 3.79 dB), and the greatest interpretability, showing improvements of 0.15 dB in global RMSE and MAE (P < 0.05) compared to the performance on original maps. Feature maps and visualization tools indicate that artifacts compromise DINO-ViT's predictive ability but improve with artifact correction.

Conclusions: Combining self-supervised ViTs with generative artifact correction enhances the correlation between glaucomatous structures and functions.

Translational relevance: Our approach offers a comprehensive tool for glaucoma management, facilitates the exploration of structure-function correlations in research, and underscores the importance of addressing artifacts in the clinical interpretation of OCT.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Translational Vision Science & Technology
Translational Vision Science & Technology Engineering-Biomedical Engineering
CiteScore
5.70
自引率
3.30%
发文量
346
审稿时长
25 weeks
期刊介绍: Translational Vision Science & Technology (TVST), an official journal of the Association for Research in Vision and Ophthalmology (ARVO), an international organization whose purpose is to advance research worldwide into understanding the visual system and preventing, treating and curing its disorders, is an online, open access, peer-reviewed journal emphasizing multidisciplinary research that bridges the gap between basic research and clinical care. A highly qualified and diverse group of Associate Editors and Editorial Board Members is led by Editor-in-Chief Marco Zarbin, MD, PhD, FARVO. The journal covers a broad spectrum of work, including but not limited to: Applications of stem cell technology for regenerative medicine, Development of new animal models of human diseases, Tissue bioengineering, Chemical engineering to improve virus-based gene delivery, Nanotechnology for drug delivery, Design and synthesis of artificial extracellular matrices, Development of a true microsurgical operating environment, Refining data analysis algorithms to improve in vivo imaging technology, Results of Phase 1 clinical trials, Reverse translational ("bedside to bench") research. TVST seeks manuscripts from scientists and clinicians with diverse backgrounds ranging from basic chemistry to ophthalmic surgery that will advance or change the way we understand and/or treat vision-threatening diseases. TVST encourages the use of color, multimedia, hyperlinks, program code and other digital enhancements.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信