Independent Evaluation of RETFound Foundation Model's Performance on Optic Nerve Analysis Using Fundus Photography

IF 3.2 Q1 OPHTHALMOLOGY
Maggie S. Chen , Rohith Ravindranath MS , Robert Chang MD , Yukun Zhou PhD , Pearse A. Keane MD FRCOphth , Sophia Y. Wang MD, MS
{"title":"Independent Evaluation of RETFound Foundation Model's Performance on Optic Nerve Analysis Using Fundus Photography","authors":"Maggie S. Chen ,&nbsp;Rohith Ravindranath MS ,&nbsp;Robert Chang MD ,&nbsp;Yukun Zhou PhD ,&nbsp;Pearse A. Keane MD FRCOphth ,&nbsp;Sophia Y. Wang MD, MS","doi":"10.1016/j.xops.2025.100720","DOIUrl":null,"url":null,"abstract":"<div><h3>Purpose</h3><div>This study evaluates RETFound, a retinal image foundation model, as a feature extractor for predicting optic nerve metrics like cup-to-disc ratio (CDR) and retinal nerve fiber layer (RNFL) thickness using an independent clinical dataset.</div></div><div><h3>Design</h3><div>Retrospective observational study.</div></div><div><h3>Participants</h3><div>Patients who underwent fundus photography and RNFL OCT at the Byers Eye Institute, Stanford University.</div></div><div><h3>Methods</h3><div>Fundus images were paired with RNFL OCT results where study dates were within 6 months of each other. Latent features from full-sized raw fundus images were extracted from RETFound and used as inputs for several linear regression models (Ridge, Lasso, Elastic Net, and ordinary least squares). Baseline models using pretrained VGG16 and Vision Transformers (ViTs) as feature extractors were also developed. All models were trained to perform single-output tasks (predicting CDR or average RNFL thickness) and multioutput tasks (predicting RNFL thickness at quadrants and clock hours). Data were split 80:20 at the patient level for training and validation.</div></div><div><h3>Main Outcome Measures</h3><div>Model predictions were evaluated on a test set using the metrics of <em>R</em><sup><em>2</em></sup>, mean absolute error, and root mean square error.</div></div><div><h3>Results</h3><div>Among the 463 unique participants, contributing 776 fundus–OCT data pairs, the mean age was 63 years (±18 years), with 57.24% being female (N = 265). RETFound models demonstrated strong performance on single-output tasks, achieving <em>R</em><sup><em>2</em></sup> values between 0.706 and 0.898 for CDR prediction and between 0.855 and 0.961 for average RNFL thickness prediction. Performance on multioutput tasks was less robust, with a highest <em>R</em><sup><em>2</em></sup> of 0.583 for clock-hour RNFL thickness prediction and an <em>R</em><sup><em>2</em></sup> of 0.811 for quadrant RNFL thickness prediction. RETFound models outperformed VGG16 and ViT models, which achieved maximum <em>R</em><sup><em>2</em></sup> of 0.731 and 0.687 in predicting RNFL thickness and CDR.</div></div><div><h3>Conclusions</h3><div>Machine learning models leveraging the massively pretrained RETFound foundation model could accurately predict CDR and average RNFL thickness from fundus photos on an independent clinical dataset. Although RETFound was not trained or fine-tuned for these optic nerve evaluation tasks, nevertheless, RETFound overcomes small dataset limitations and excels in specialized applications.</div></div><div><h3>Financial Disclosure(s)</h3><div>Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.</div></div>","PeriodicalId":74363,"journal":{"name":"Ophthalmology science","volume":"5 3","pages":"Article 100720"},"PeriodicalIF":3.2000,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ophthalmology science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666914525000181","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose

This study evaluates RETFound, a retinal image foundation model, as a feature extractor for predicting optic nerve metrics like cup-to-disc ratio (CDR) and retinal nerve fiber layer (RNFL) thickness using an independent clinical dataset.

Design

Retrospective observational study.

Participants

Patients who underwent fundus photography and RNFL OCT at the Byers Eye Institute, Stanford University.

Methods

Fundus images were paired with RNFL OCT results where study dates were within 6 months of each other. Latent features from full-sized raw fundus images were extracted from RETFound and used as inputs for several linear regression models (Ridge, Lasso, Elastic Net, and ordinary least squares). Baseline models using pretrained VGG16 and Vision Transformers (ViTs) as feature extractors were also developed. All models were trained to perform single-output tasks (predicting CDR or average RNFL thickness) and multioutput tasks (predicting RNFL thickness at quadrants and clock hours). Data were split 80:20 at the patient level for training and validation.

Main Outcome Measures

Model predictions were evaluated on a test set using the metrics of R2, mean absolute error, and root mean square error.

Results

Among the 463 unique participants, contributing 776 fundus–OCT data pairs, the mean age was 63 years (±18 years), with 57.24% being female (N = 265). RETFound models demonstrated strong performance on single-output tasks, achieving R2 values between 0.706 and 0.898 for CDR prediction and between 0.855 and 0.961 for average RNFL thickness prediction. Performance on multioutput tasks was less robust, with a highest R2 of 0.583 for clock-hour RNFL thickness prediction and an R2 of 0.811 for quadrant RNFL thickness prediction. RETFound models outperformed VGG16 and ViT models, which achieved maximum R2 of 0.731 and 0.687 in predicting RNFL thickness and CDR.

Conclusions

Machine learning models leveraging the massively pretrained RETFound foundation model could accurately predict CDR and average RNFL thickness from fundus photos on an independent clinical dataset. Although RETFound was not trained or fine-tuned for these optic nerve evaluation tasks, nevertheless, RETFound overcomes small dataset limitations and excels in specialized applications.

Financial Disclosure(s)

Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
RETFound基础模型在眼底摄影视神经分析中的独立评价
目的:本研究使用独立的临床数据集,评估RETFound视网膜图像基础模型作为预测视神经指标(如杯盘比(CDR)和视网膜神经纤维层(RNFL)厚度)的特征提取器。设计回顾性观察性研究。在斯坦福大学拜尔斯眼科研究所接受眼底摄影和RNFL OCT的患者。方法眼底图像与RNFL OCT结果配对,研究日期在6个月内。从RETFound中提取全尺寸原始眼底图像的潜在特征,并将其用作几种线性回归模型(Ridge, Lasso, Elastic Net和普通最小二乘法)的输入。使用预训练的VGG16和视觉变压器(ViTs)作为特征提取器的基线模型也被开发出来。所有模型都经过训练,可以执行单输出任务(预测CDR或平均RNFL厚度)和多输出任务(预测象限和时钟小时的RNFL厚度)。为了训练和验证,数据在患者水平上被分割为80:20。主要结局测量采用R2、平均绝对误差和均方根误差等指标在测试集上对模型预测进行评估。结果463例独特参与者提供776对眼底oct资料,平均年龄63岁(±18岁),女性占57.24% (N = 265)。RETFound模型在单输出任务上表现出色,CDR预测的R2值在0.706 ~ 0.898之间,平均RNFL厚度预测的R2值在0.855 ~ 0.961之间。多输出任务的性能较差,时钟小时RNFL厚度预测的R2最高为0.583,象限RNFL厚度预测的R2最高为0.811。RETFound模型预测RNFL厚度和CDR的R2分别为0.731和0.687,优于VGG16和ViT模型。结论利用大规模预训练RETFound基础模型的机器学习模型可以准确预测独立临床数据集眼底照片的CDR和平均RNFL厚度。尽管RETFound没有针对这些视神经评估任务进行训练或微调,但是,RETFound克服了小数据集的限制,并在专业应用中表现出色。财务披露专有或商业披露可在本文末尾的脚注和披露中找到。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Ophthalmology science
Ophthalmology science Ophthalmology
CiteScore
3.40
自引率
0.00%
发文量
0
审稿时长
89 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信