From Image to Sequence: Exploring Vision Transformers for Optical Coherence Tomography Classification.

IF 1.1 Q4 ENGINEERING, BIOMEDICAL
Journal of Medical Signals & Sensors Pub Date : 2025-06-09 eCollection Date: 2025-01-01 DOI:10.4103/jmss.jmss_58_24
Amirali Arbab, Aref Habibi, Hossein Rabbani, Mahnoosh Tajmirriahi
{"title":"From Image to Sequence: Exploring Vision Transformers for Optical Coherence Tomography Classification.","authors":"Amirali Arbab, Aref Habibi, Hossein Rabbani, Mahnoosh Tajmirriahi","doi":"10.4103/jmss.jmss_58_24","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Optical coherence tomography (OCT) is a pivotal imaging technique for the early detection and management of critical retinal diseases, notably diabetic macular edema and age-related macular degeneration. These conditions are significant global health concerns, affecting millions and leading to vision loss if not diagnosed promptly. Current methods for OCT image classification encounter specific challenges, such as the inherent complexity of retinal structures and considerable variability across different OCT datasets.</p><p><strong>Methods: </strong>This paper introduces a novel hybrid model that integrates the strengths of convolutional neural networks (CNNs) and vision transformer (ViT) to overcome these obstacles. The synergy between CNNs, which excel at extracting detailed localized features, and ViT, adept at recognizing long-range patterns, enables a more effective and comprehensive analysis of OCT images.</p><p><strong>Results: </strong>While our model achieves an accuracy of 99.80% on the OCT2017 dataset, its standout feature is its parameter efficiency-requiring only 6.9 million parameters, significantly fewer than larger, more complex models such as Xception and OpticNet-71.</p><p><strong>Conclusion: </strong>This efficiency underscores the model's suitability for clinical settings, where computational resources may be limited but high accuracy and rapid diagnosis are imperative.<b>Code Availability:</b> The code for this study is available at https://github.com/Amir1831/ViT4OCT.</p>","PeriodicalId":37680,"journal":{"name":"Journal of Medical Signals & Sensors","volume":"15 ","pages":"18"},"PeriodicalIF":1.1000,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12180780/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Medical Signals & Sensors","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4103/jmss.jmss_58_24","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q4","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Optical coherence tomography (OCT) is a pivotal imaging technique for the early detection and management of critical retinal diseases, notably diabetic macular edema and age-related macular degeneration. These conditions are significant global health concerns, affecting millions and leading to vision loss if not diagnosed promptly. Current methods for OCT image classification encounter specific challenges, such as the inherent complexity of retinal structures and considerable variability across different OCT datasets.

Methods: This paper introduces a novel hybrid model that integrates the strengths of convolutional neural networks (CNNs) and vision transformer (ViT) to overcome these obstacles. The synergy between CNNs, which excel at extracting detailed localized features, and ViT, adept at recognizing long-range patterns, enables a more effective and comprehensive analysis of OCT images.

Results: While our model achieves an accuracy of 99.80% on the OCT2017 dataset, its standout feature is its parameter efficiency-requiring only 6.9 million parameters, significantly fewer than larger, more complex models such as Xception and OpticNet-71.

Conclusion: This efficiency underscores the model's suitability for clinical settings, where computational resources may be limited but high accuracy and rapid diagnosis are imperative.Code Availability: The code for this study is available at https://github.com/Amir1831/ViT4OCT.

从图像到序列:探索光学相干层析成像分类的视觉变换。
背景:光学相干断层扫描(OCT)是早期发现和治疗关键视网膜疾病的关键成像技术,特别是糖尿病性黄斑水肿和年龄相关性黄斑变性。这些疾病是全球重大的健康问题,影响数百万人,如果不及时诊断,会导致视力丧失。当前的OCT图像分类方法遇到了特定的挑战,例如视网膜结构的固有复杂性和不同OCT数据集的相当大的可变性。方法:引入一种新的混合模型,将卷积神经网络(cnn)和视觉变压器(ViT)的优势结合起来,克服这些障碍。擅长提取细节局部特征的cnn和擅长识别远程模式的ViT的协同作用,使OCT图像的分析更加有效和全面。结果:虽然我们的模型在OCT2017数据集上达到了99.80%的准确率,但其突出的特点是参数效率——只需要690万个参数,比Xception和OpticNet-71等更大、更复杂的模型要少得多。结论:这种效率强调了该模型对临床环境的适用性,在临床环境中,计算资源可能有限,但高精度和快速诊断是必不可少的。代码可用性:本研究的代码可在https://github.com/Amir1831/ViT4OCT上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Medical Signals & Sensors
Journal of Medical Signals & Sensors ENGINEERING, BIOMEDICAL-
CiteScore
2.30
自引率
0.00%
发文量
53
审稿时长
33 weeks
期刊介绍: JMSS is an interdisciplinary journal that incorporates all aspects of the biomedical engineering including bioelectrics, bioinformatics, medical physics, health technology assessment, etc. Subject areas covered by the journal include: - Bioelectric: Bioinstruments Biosensors Modeling Biomedical signal processing Medical image analysis and processing Medical imaging devices Control of biological systems Neuromuscular systems Cognitive sciences Telemedicine Robotic Medical ultrasonography Bioelectromagnetics Electrophysiology Cell tracking - Bioinformatics and medical informatics: Analysis of biological data Data mining Stochastic modeling Computational genomics Artificial intelligence & fuzzy Applications Medical softwares Bioalgorithms Electronic health - Biophysics and medical physics: Computed tomography Radiation therapy Laser therapy - Education in biomedical engineering - Health technology assessment - Standard in biomedical engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信