Saif Khalid, Hatem A. Rashwan, Saddam Abdulwahab, Mohamed Abdel-Nasser, Facundo Manuel Quiroga, Domenec Puig
{"title":"FGR-Net:Interpretable fundus imagegradeability classification based on deepreconstruction learning","authors":"Saif Khalid, Hatem A. Rashwan, Saddam Abdulwahab, Mohamed Abdel-Nasser, Facundo Manuel Quiroga, Domenec Puig","doi":"arxiv-2409.10246","DOIUrl":null,"url":null,"abstract":"The performance of diagnostic Computer-Aided Design (CAD) systems for retinal\ndiseases depends on the quality of the retinal images being screened. Thus,\nmany studies have been developed to evaluate and assess the quality of such\nretinal images. However, most of them did not investigate the relationship\nbetween the accuracy of the developed models and the quality of the\nvisualization of interpretability methods for distinguishing between gradable\nand non-gradable retinal images. Consequently, this paper presents a novel\nframework called FGR-Net to automatically assess and interpret underlying\nfundus image quality by merging an autoencoder network with a classifier\nnetwork. The FGR-Net model also provides an interpretable quality assessment\nthrough visualizations. In particular, FGR-Net uses a deep autoencoder to\nreconstruct the input image in order to extract the visual characteristics of\nthe input fundus images based on self-supervised learning. The extracted\nfeatures by the autoencoder are then fed into a deep classifier network to\ndistinguish between gradable and ungradable fundus images. FGR-Net is evaluated\nwith different interpretability methods, which indicates that the autoencoder\nis a key factor in forcing the classifier to focus on the relevant structures\nof the fundus images, such as the fovea, optic disk, and prominent blood\nvessels. Additionally, the interpretability methods can provide visual feedback\nfor ophthalmologists to understand how our model evaluates the quality of\nfundus images. The experimental results showed the superiority of FGR-Net over\nthe state-of-the-art quality assessment methods, with an accuracy of 89% and an\nF1-score of 87%.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":"18 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Image and Video Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10246","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The performance of diagnostic Computer-Aided Design (CAD) systems for retinal
diseases depends on the quality of the retinal images being screened. Thus,
many studies have been developed to evaluate and assess the quality of such
retinal images. However, most of them did not investigate the relationship
between the accuracy of the developed models and the quality of the
visualization of interpretability methods for distinguishing between gradable
and non-gradable retinal images. Consequently, this paper presents a novel
framework called FGR-Net to automatically assess and interpret underlying
fundus image quality by merging an autoencoder network with a classifier
network. The FGR-Net model also provides an interpretable quality assessment
through visualizations. In particular, FGR-Net uses a deep autoencoder to
reconstruct the input image in order to extract the visual characteristics of
the input fundus images based on self-supervised learning. The extracted
features by the autoencoder are then fed into a deep classifier network to
distinguish between gradable and ungradable fundus images. FGR-Net is evaluated
with different interpretability methods, which indicates that the autoencoder
is a key factor in forcing the classifier to focus on the relevant structures
of the fundus images, such as the fovea, optic disk, and prominent blood
vessels. Additionally, the interpretability methods can provide visual feedback
for ophthalmologists to understand how our model evaluates the quality of
fundus images. The experimental results showed the superiority of FGR-Net over
the state-of-the-art quality assessment methods, with an accuracy of 89% and an
F1-score of 87%.