Satyabrata Lenka , Zefree Lazarus Mayaluri , Ganapati Panda
{"title":"Glaucoma detection from retinal fundus images using graph convolution based multi-task model","authors":"Satyabrata Lenka , Zefree Lazarus Mayaluri , Ganapati Panda","doi":"10.1016/j.prime.2025.100931","DOIUrl":null,"url":null,"abstract":"<div><div>Glaucoma is an abnormality in the eye condition that, if not treated within a safe time limit, can result in visual loss. Glaucoma diagnosis requires professionals to identify minor structural changes in the structure of the optic disc and optic cup from retinal fundus images in a short period. Deep learning algorithms have been employed effectively in the segmentation of biomedical images and the identification of diseases. To accomplish good generalization, model training requires comprehensive annotations, which is a difficult task. The intended objective of the present study is to come up with and train a distinctive multi-task deep learning model for automated fundus image segmentation and classification. The multi-task model learns for the segmentation task of Optic Disc (OD) and Optic Cup (OC) and the classification task for accurate glaucoma detection using both structural and image-based features. The multi-task model proposed a modified U-net architecture in which Mobile-Netv2 is used in the encoder part, Graph Convolution Network (GCN) is used in the decoder part, and an attention module (AM) is used to locate the region of interest (ROI) for better feature extraction. The implementation of this model is done using three fundus image datasets such as ORIGA, REFUGE, and DRISTI-GS. The performance of the proposed multi-task model is compared with some existing methods. It shows maximum accuracy of 97<em>.</em>43 % and AUROC of 0.985 for the glaucoma detection task and high-quality OD and OC segmented images with dice coefficient of 97<em>.</em>95 % and 96<em>.</em>11 % respectively for the segmentation task.</div></div>","PeriodicalId":100488,"journal":{"name":"e-Prime - Advances in Electrical Engineering, Electronics and Energy","volume":"11 ","pages":"Article 100931"},"PeriodicalIF":0.0000,"publicationDate":"2025-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"e-Prime - Advances in Electrical Engineering, Electronics and Energy","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772671125000385","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Glaucoma is an abnormality in the eye condition that, if not treated within a safe time limit, can result in visual loss. Glaucoma diagnosis requires professionals to identify minor structural changes in the structure of the optic disc and optic cup from retinal fundus images in a short period. Deep learning algorithms have been employed effectively in the segmentation of biomedical images and the identification of diseases. To accomplish good generalization, model training requires comprehensive annotations, which is a difficult task. The intended objective of the present study is to come up with and train a distinctive multi-task deep learning model for automated fundus image segmentation and classification. The multi-task model learns for the segmentation task of Optic Disc (OD) and Optic Cup (OC) and the classification task for accurate glaucoma detection using both structural and image-based features. The multi-task model proposed a modified U-net architecture in which Mobile-Netv2 is used in the encoder part, Graph Convolution Network (GCN) is used in the decoder part, and an attention module (AM) is used to locate the region of interest (ROI) for better feature extraction. The implementation of this model is done using three fundus image datasets such as ORIGA, REFUGE, and DRISTI-GS. The performance of the proposed multi-task model is compared with some existing methods. It shows maximum accuracy of 97.43 % and AUROC of 0.985 for the glaucoma detection task and high-quality OD and OC segmented images with dice coefficient of 97.95 % and 96.11 % respectively for the segmentation task.