Samantha Min Er Yew BSc , Xiaofeng Lei MSc , Yibing Chen BEng , Jocelyn Hui Lin Goh BEng , Krithi Pushpanathan MSc , Can Can Xue MD, PhD , Ya Xing Wang MD, PhD , Jost B. Jonas MD, PhD , Charumathi Sabanayagam MD, PhD , Victor Teck Chang Koh MBBS, MMed , Xinxing Xu PhD , Yong Liu PhD , Ching-Yu Cheng MD, PhD , Yih-Chung Tham PhD
{"title":"Deep Imbalanced Regression Model for Predicting Refractive Error from Retinal Photos","authors":"Samantha Min Er Yew BSc , Xiaofeng Lei MSc , Yibing Chen BEng , Jocelyn Hui Lin Goh BEng , Krithi Pushpanathan MSc , Can Can Xue MD, PhD , Ya Xing Wang MD, PhD , Jost B. Jonas MD, PhD , Charumathi Sabanayagam MD, PhD , Victor Teck Chang Koh MBBS, MMed , Xinxing Xu PhD , Yong Liu PhD , Ching-Yu Cheng MD, PhD , Yih-Chung Tham PhD","doi":"10.1016/j.xops.2024.100659","DOIUrl":null,"url":null,"abstract":"<div><h3>Purpose</h3><div>Recent studies utilized ocular images and deep learning (DL) to predict refractive error and yielded notable results. However, most studies did not address biases from imbalanced datasets or conduct external validations. To address these gaps, this study aimed to integrate the deep imbalanced regression (DIR) technique into ResNet and Vision Transformer models to predict refractive error from retinal photographs.</div></div><div><h3>Design</h3><div>Retrospective study.</div></div><div><h3>Subjects</h3><div>We developed the DL models using up to 103 865 images from the Singapore Epidemiology of Eye Diseases Study and the United Kingdom Biobank, with internal testing on up to 8067 images. External testing was conducted on 7043 images from the Singapore Prospective Study and 5539 images from the Beijing Eye Study. Retinal images and corresponding refractive error data were extracted.</div></div><div><h3>Methods</h3><div>This retrospective study developed regression-based models, including ResNet34 with DIR, and SwinV2 (Swin Transformer) with DIR, incorporating Label Distribution Smoothing and Feature Distribution Smoothing. These models were compared against their baseline versions, ResNet34 and SwinV2, in predicting spherical and spherical equivalent (SE) power.</div></div><div><h3>Main Outcome Measures</h3><div>Mean absolute error (MAE) and coefficient of determination were used to evaluate the models’ performances. The Wilcoxon signed-rank test was performed to assess statistical significance between DIR-integrated models and their baseline versions.</div></div><div><h3>Results</h3><div>For prediction of the spherical power, ResNet34 with DIR (MAE: 0.84D) and SwinV2 with DIR (MAE: 0.77D) significantly outperformed their baseline—ResNet34 (MAE: 0.88D; <em>P</em> < 0.001) and SwinV2 (MAE: 0.87D; <em>P</em> < 0.001) in internal test. For prediction of the SE power, ResNet34 with DIR (MAE: 0.78D) and SwinV2 with DIR (MAE: 0.75D) consistently significantly outperformed its baseline—ResNet34 (MAE: 0.81D; <em>P</em> < 0.001) and SwinV2 (MAE: 0.78D; <em>P</em> < 0.05) in internal test. Similar trends were observed in external test sets for both spherical and SE power prediction.</div></div><div><h3>Conclusions</h3><div>Deep imbalanced regressed–integrated DL models showed potential in addressing data imbalances and improving the prediction of refractive error. These findings highlight the potential utility of combining DL models with retinal imaging for opportunistic screening of refractive errors, particularly in settings where retinal cameras are already in use.</div></div><div><h3>Financial Disclosure(s)</h3><div>Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.</div></div>","PeriodicalId":74363,"journal":{"name":"Ophthalmology science","volume":"5 2","pages":"Article 100659"},"PeriodicalIF":3.2000,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ophthalmology science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666914524001957","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose
Recent studies utilized ocular images and deep learning (DL) to predict refractive error and yielded notable results. However, most studies did not address biases from imbalanced datasets or conduct external validations. To address these gaps, this study aimed to integrate the deep imbalanced regression (DIR) technique into ResNet and Vision Transformer models to predict refractive error from retinal photographs.
Design
Retrospective study.
Subjects
We developed the DL models using up to 103 865 images from the Singapore Epidemiology of Eye Diseases Study and the United Kingdom Biobank, with internal testing on up to 8067 images. External testing was conducted on 7043 images from the Singapore Prospective Study and 5539 images from the Beijing Eye Study. Retinal images and corresponding refractive error data were extracted.
Methods
This retrospective study developed regression-based models, including ResNet34 with DIR, and SwinV2 (Swin Transformer) with DIR, incorporating Label Distribution Smoothing and Feature Distribution Smoothing. These models were compared against their baseline versions, ResNet34 and SwinV2, in predicting spherical and spherical equivalent (SE) power.
Main Outcome Measures
Mean absolute error (MAE) and coefficient of determination were used to evaluate the models’ performances. The Wilcoxon signed-rank test was performed to assess statistical significance between DIR-integrated models and their baseline versions.
Results
For prediction of the spherical power, ResNet34 with DIR (MAE: 0.84D) and SwinV2 with DIR (MAE: 0.77D) significantly outperformed their baseline—ResNet34 (MAE: 0.88D; P < 0.001) and SwinV2 (MAE: 0.87D; P < 0.001) in internal test. For prediction of the SE power, ResNet34 with DIR (MAE: 0.78D) and SwinV2 with DIR (MAE: 0.75D) consistently significantly outperformed its baseline—ResNet34 (MAE: 0.81D; P < 0.001) and SwinV2 (MAE: 0.78D; P < 0.05) in internal test. Similar trends were observed in external test sets for both spherical and SE power prediction.
Conclusions
Deep imbalanced regressed–integrated DL models showed potential in addressing data imbalances and improving the prediction of refractive error. These findings highlight the potential utility of combining DL models with retinal imaging for opportunistic screening of refractive errors, particularly in settings where retinal cameras are already in use.
Financial Disclosure(s)
Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.