{"title":"Glaucoformer: Dual-domain Global Transformer Network for Generalized Glaucoma Stage Classification.","authors":"Dipankar Das, Deepak Ranjan Nayak, Ram Bilas Pachori","doi":"10.1109/JBHI.2025.3574997","DOIUrl":null,"url":null,"abstract":"<p><p>Classification of glaucoma stages remains challenging due to substantial inter-stage similarities, the presence of irrelevant features, and subtle lesion size, shape, and color variations in fundus images. For this purpose, few efforts have recently been made using traditional machine learning and deep learning models, specifically convolutional neural networks (CNN). While the conventional CNN models capture local contextual features within fixed receptive fields, they fail to exploit global contextual dependencies. Transformers, on the other hand, are capable of modeling global contextual information. However, they lack the ability to capture local contexts and merely focus on performing attention in the spatial domain, ignoring feature analysis in the frequency domain. To address these issues, we present a novel dual-domain global transformer network, Glaucoformer, to effectively classify glaucoma stages. Specifically, we propose a dual-domain global transformer layer (DGTL) consisting of dual-domain channel attention (DCA) and dual-domain spatial attention (DSA) with Fourier domain feature analyzer (FDFA) as the core component and integrated with a backbone. This helps in exploiting local and global contextual feature dependencies in both spatial and frequency domains, thereby learning prominent and discriminant feature representations. A shared key-query scheme is introduced to learn complementary features while reducing the parameters. In addition, the DGTL leverages the benefits of a deformable convolution to enable the model to handle complex lesion irregularities. We evaluate our method on a benchmark dataset, and the experimental results and extensive comparisons with existing CNN and vision transformer-based approaches indicate its effectiveness for glaucoma stage classification. Also, the results on an unseen dataset demonstrate the generalizability of the model.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7000,"publicationDate":"2025-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Biomedical and Health Informatics","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1109/JBHI.2025.3574997","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Classification of glaucoma stages remains challenging due to substantial inter-stage similarities, the presence of irrelevant features, and subtle lesion size, shape, and color variations in fundus images. For this purpose, few efforts have recently been made using traditional machine learning and deep learning models, specifically convolutional neural networks (CNN). While the conventional CNN models capture local contextual features within fixed receptive fields, they fail to exploit global contextual dependencies. Transformers, on the other hand, are capable of modeling global contextual information. However, they lack the ability to capture local contexts and merely focus on performing attention in the spatial domain, ignoring feature analysis in the frequency domain. To address these issues, we present a novel dual-domain global transformer network, Glaucoformer, to effectively classify glaucoma stages. Specifically, we propose a dual-domain global transformer layer (DGTL) consisting of dual-domain channel attention (DCA) and dual-domain spatial attention (DSA) with Fourier domain feature analyzer (FDFA) as the core component and integrated with a backbone. This helps in exploiting local and global contextual feature dependencies in both spatial and frequency domains, thereby learning prominent and discriminant feature representations. A shared key-query scheme is introduced to learn complementary features while reducing the parameters. In addition, the DGTL leverages the benefits of a deformable convolution to enable the model to handle complex lesion irregularities. We evaluate our method on a benchmark dataset, and the experimental results and extensive comparisons with existing CNN and vision transformer-based approaches indicate its effectiveness for glaucoma stage classification. Also, the results on an unseen dataset demonstrate the generalizability of the model.
期刊介绍:
IEEE Journal of Biomedical and Health Informatics publishes original papers presenting recent advances where information and communication technologies intersect with health, healthcare, life sciences, and biomedicine. Topics include acquisition, transmission, storage, retrieval, management, and analysis of biomedical and health information. The journal covers applications of information technologies in healthcare, patient monitoring, preventive care, early disease diagnosis, therapy discovery, and personalized treatment protocols. It explores electronic medical and health records, clinical information systems, decision support systems, medical and biological imaging informatics, wearable systems, body area/sensor networks, and more. Integration-related topics like interoperability, evidence-based medicine, and secure patient data are also addressed.