Utilization of Image-Based Deep Learning in Multimodal Glaucoma Detection Neural Network from a Primary Patient Cohort

IF 3.2 Q1 OPHTHALMOLOGY
Elizabeth E. Hwang PhD , Dake Chen PhD , Ying Han MD, PhD , Lin Jia PhD , Jing Shan MD, PhD
{"title":"Utilization of Image-Based Deep Learning in Multimodal Glaucoma Detection Neural Network from a Primary Patient Cohort","authors":"Elizabeth E. Hwang PhD ,&nbsp;Dake Chen PhD ,&nbsp;Ying Han MD, PhD ,&nbsp;Lin Jia PhD ,&nbsp;Jing Shan MD, PhD","doi":"10.1016/j.xops.2025.100703","DOIUrl":null,"url":null,"abstract":"<div><h3>Purpose</h3><div>To develop a clinically motivated multimodal neural network <u>g</u>laucoma detection model trained on minimally processed imaging data of time-matched multimodal testing including fundus photographs, OCT scans, and Humphrey visual field (HVF) analysis.</div></div><div><h3>Design</h3><div>Evaluation of a diagnostic technology.</div></div><div><h3>Subjects</h3><div>A total of 716 encounters with time-matched fundus photographs, OCT optic nerve imaging, and HVF testing from 706 eyes (557 nonglaucomatous, 149 glaucomatous) from 571 individual patients seen at a tertiary medical center and 4 external single-modality (fundus photograph and OCT) datasets.</div></div><div><h3>Methods</h3><div>A multimodal neural network model was developed consisting of 2 main components: first, 3 convolutional neural networks to extract semantic features and generate embeddings for each respective modality, followed by a second component consisting of a multilayer perceptron to integrate the individual embeddings and produce a predicted label, glaucomatous or nonglaucomatous.</div></div><div><h3>Main Outcome Measures</h3><div>Single and multimodal performances were evaluated on the internal test set using the area under the receiver operating characteristic curve (AUC), accuracy, recall, and specificity. Fundus photograph and OCT single-modality neural networks were additionally evaluated on external datasets by these metrics.</div></div><div><h3>Results</h3><div>Our results show single-modality models with high performance on curated training datasets perform inferiorly on our primary clinical dataset. Performance metrics, however, can be notably improved through multimodal integration (AUC: 0.86 from 0.57 to 0.74 and specificity: 0.85 from 0.77 to 0.82), suggesting that a holistic approach considering both structural and functional data may enhance the functionality and accuracy of artificial intelligence (AI) model.</div></div><div><h3>Conclusions</h3><div>Clinical implementation of deep learning models for glaucoma detection benefits from multimodal integration, and we demonstrate this approach on a true clinical cohort to obtain a production-level AI solution for glaucoma diagnosis.</div></div><div><h3>Financial Disclosure(s)</h3><div>Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.</div></div>","PeriodicalId":74363,"journal":{"name":"Ophthalmology science","volume":"5 3","pages":"Article 100703"},"PeriodicalIF":3.2000,"publicationDate":"2025-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ophthalmology science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666914525000016","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose

To develop a clinically motivated multimodal neural network glaucoma detection model trained on minimally processed imaging data of time-matched multimodal testing including fundus photographs, OCT scans, and Humphrey visual field (HVF) analysis.

Design

Evaluation of a diagnostic technology.

Subjects

A total of 716 encounters with time-matched fundus photographs, OCT optic nerve imaging, and HVF testing from 706 eyes (557 nonglaucomatous, 149 glaucomatous) from 571 individual patients seen at a tertiary medical center and 4 external single-modality (fundus photograph and OCT) datasets.

Methods

A multimodal neural network model was developed consisting of 2 main components: first, 3 convolutional neural networks to extract semantic features and generate embeddings for each respective modality, followed by a second component consisting of a multilayer perceptron to integrate the individual embeddings and produce a predicted label, glaucomatous or nonglaucomatous.

Main Outcome Measures

Single and multimodal performances were evaluated on the internal test set using the area under the receiver operating characteristic curve (AUC), accuracy, recall, and specificity. Fundus photograph and OCT single-modality neural networks were additionally evaluated on external datasets by these metrics.

Results

Our results show single-modality models with high performance on curated training datasets perform inferiorly on our primary clinical dataset. Performance metrics, however, can be notably improved through multimodal integration (AUC: 0.86 from 0.57 to 0.74 and specificity: 0.85 from 0.77 to 0.82), suggesting that a holistic approach considering both structural and functional data may enhance the functionality and accuracy of artificial intelligence (AI) model.

Conclusions

Clinical implementation of deep learning models for glaucoma detection benefits from multimodal integration, and we demonstrate this approach on a true clinical cohort to obtain a production-level AI solution for glaucoma diagnosis.

Financial Disclosure(s)

Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
基于图像的深度学习在多模态青光眼检测神经网络中的应用
目的建立一种临床驱动的多模态神经网络青光眼检测模型,该模型基于时间匹配多模态检测的最小处理成像数据,包括眼底照片、OCT扫描和Humphrey视野(HVF)分析。一种诊断技术的设计评估。在三级医疗中心和4个外部单模态(眼底照片和OCT)数据集就诊的571例患者的706只眼睛(557只为非青光眼,149只为青光眼)共716次眼底照片、OCT视神经成像和HVF测试。方法建立一个多模态神经网络模型,该模型由2个主要组成部分组成:首先,3个卷积神经网络提取语义特征并为每个模态生成嵌入;然后,由多层感知器组成的第二部分整合各个嵌入并产生青光眼或非青光眼的预测标签。在内部测试集上使用受试者工作特征曲线下面积(AUC)、准确性、召回率和特异性评估单模式和多模式性能。眼底照片和OCT单模态神经网络通过这些指标在外部数据集上进行额外评估。我们的研究结果表明,在精心策划的训练数据集上表现优异的单模态模型在我们的主要临床数据集上表现较差。然而,通过多模态集成可以显著改善性能指标(AUC从0.57到0.74:0.86,特异性从0.77到0.82:0.85),这表明考虑结构和功能数据的整体方法可以增强人工智能(AI)模型的功能和准确性。青光眼检测的深度学习模型的临床实施得益于多模式集成,我们在一个真正的临床队列中展示了这种方法,以获得青光眼诊断的生产级人工智能解决方案。财务披露专有或商业披露可在本文末尾的脚注和披露中找到。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Ophthalmology science
Ophthalmology science Ophthalmology
CiteScore
3.40
自引率
0.00%
发文量
0
审稿时长
89 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信