Feng Yan , Yanxia Liu , Qingsong Zhao , Guangguo He
{"title":"A multi-task deep learning pipeline integrating vessel segmentation and radiomics for multiclass retinal disease classification","authors":"Feng Yan , Yanxia Liu , Qingsong Zhao , Guangguo He","doi":"10.1016/j.pdpdt.2025.105209","DOIUrl":null,"url":null,"abstract":"<div><h3>Objective</h3><div>This study aims to develop a robust, multi-task deep learning framework that integrates vessel segmentation and radiomic analysis for the automated classification of four retinal conditions— diabetic retinopathy (DR), hypertensive retinopathy (HR), papilledema, and normal fundus—using fundus images.</div></div><div><h3>Materials and Methods</h3><div>A total of 2165 patients from eight medical centers were enrolled. Fundus images underwent standardized preprocessing including histogram equalization, normalization, resizing, and augmentation. Whole vessel and artery-vein segmentations were conducted using five deep learning models: U-Net, Attention U-Net, DeepLabV3+, HRNet, and Swin-Unet. From the segmented vascular maps, 220 radiomic features were extracted using PyRadiomics and Mahotas toolkits. The arteriovenous ratio (AVR) was also computed from artery and vein masks. ICC analysis was used to assess reproducibility across centers, with features below ICC < 0.75 excluded. Feature selection was performed using Least Absolute Shrinkage and Selection Operator (LASSO), Recursive Feature Elimination (RFE), and Mutual Information (MI). The combined AVR and radiomic features were input into four classifiers— Extreme Gradient Boosting (XGBoost), Categorical Boosting (CatBoost), Random Forest (RF), and Ensemble. Models were trained and validated on stratified splits and externally tested on an independent cohort of 769 patients. Evaluation metrics included accuracy, area under curve (AUC), recall, and receiver operating characteristics (ROC) analysis.</div></div><div><h3>Results</h3><div>Swin-Unet outperformed all models with external Dice Similarity Coefficient (DSC) of 92.4 % for whole vessel and 89.8 % for artery-vein segmentation. Classification using the LASSO-Ensemble combination achieved test accuracy of 93.7 %, external test accuracy of 92.3 %, and AUC of 95.2 %. AVR estimates were consistent with clinical expectations and contributed significantly to class discrimination.</div></div><div><h3>Conclusion</h3><div>This multi-task pipeline demonstrates the potential of combining transformer-based segmentation with radiomics for accurate, interpretable retinal disease classification, showing strong generalizability for future clinical applications.</div></div>","PeriodicalId":20141,"journal":{"name":"Photodiagnosis and Photodynamic Therapy","volume":"56 ","pages":"Article 105209"},"PeriodicalIF":2.6000,"publicationDate":"2025-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Photodiagnosis and Photodynamic Therapy","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1572100025007409","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ONCOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Objective
This study aims to develop a robust, multi-task deep learning framework that integrates vessel segmentation and radiomic analysis for the automated classification of four retinal conditions— diabetic retinopathy (DR), hypertensive retinopathy (HR), papilledema, and normal fundus—using fundus images.
Materials and Methods
A total of 2165 patients from eight medical centers were enrolled. Fundus images underwent standardized preprocessing including histogram equalization, normalization, resizing, and augmentation. Whole vessel and artery-vein segmentations were conducted using five deep learning models: U-Net, Attention U-Net, DeepLabV3+, HRNet, and Swin-Unet. From the segmented vascular maps, 220 radiomic features were extracted using PyRadiomics and Mahotas toolkits. The arteriovenous ratio (AVR) was also computed from artery and vein masks. ICC analysis was used to assess reproducibility across centers, with features below ICC < 0.75 excluded. Feature selection was performed using Least Absolute Shrinkage and Selection Operator (LASSO), Recursive Feature Elimination (RFE), and Mutual Information (MI). The combined AVR and radiomic features were input into four classifiers— Extreme Gradient Boosting (XGBoost), Categorical Boosting (CatBoost), Random Forest (RF), and Ensemble. Models were trained and validated on stratified splits and externally tested on an independent cohort of 769 patients. Evaluation metrics included accuracy, area under curve (AUC), recall, and receiver operating characteristics (ROC) analysis.
Results
Swin-Unet outperformed all models with external Dice Similarity Coefficient (DSC) of 92.4 % for whole vessel and 89.8 % for artery-vein segmentation. Classification using the LASSO-Ensemble combination achieved test accuracy of 93.7 %, external test accuracy of 92.3 %, and AUC of 95.2 %. AVR estimates were consistent with clinical expectations and contributed significantly to class discrimination.
Conclusion
This multi-task pipeline demonstrates the potential of combining transformer-based segmentation with radiomics for accurate, interpretable retinal disease classification, showing strong generalizability for future clinical applications.
期刊介绍:
Photodiagnosis and Photodynamic Therapy is an international journal for the dissemination of scientific knowledge and clinical developments of Photodiagnosis and Photodynamic Therapy in all medical specialties. The journal publishes original articles, review articles, case presentations, "how-to-do-it" articles, Letters to the Editor, short communications and relevant images with short descriptions. All submitted material is subject to a strict peer-review process.