Self-supervised contrastive learning improves machine learning discrimination of full thickness macular holes from epiretinal membranes in retinal OCT scans
Tim Wheeler, Kaitlyn Hunter, Patricia Anne Garcia, Henry Li, Andrew Thomson, Allan Hunter, Courosh Mehanian
{"title":"Self-supervised contrastive learning improves machine learning discrimination of full thickness macular holes from epiretinal membranes in retinal OCT scans","authors":"Tim Wheeler, Kaitlyn Hunter, Patricia Anne Garcia, Henry Li, Andrew Thomson, Allan Hunter, Courosh Mehanian","doi":"10.1101/2023.11.14.23298513","DOIUrl":null,"url":null,"abstract":"There is a growing interest in using computer-assisted models for the detection of macular conditions using optical coherence tomography (OCT) data. As the quantity of clinical scan data of specific conditions is limited, these models are typically developed by fine-tuning a generalized network to classify specific macular conditions of interest. Full thickness macular holes (FTMH) present a condition requiring timely surgical intervention to prevent permanent vision loss. Other works on automated FTMH classification have tended to use supervised ImageNet pre-trained networks with good results but leave room for improvement. In this paper, we develop a model for FTMH classification using OCT slices around the central foveal region to pre-train a naïve network using contrastive self-supervised learning. We found that self-supervised pre-trained networks outperform ImageNet pre-trained networks despite a small training set size (284 eyes total, 51 FTMH+ eyes, 3 slices from each eye). 3D spatial contrast pre-training yields a model with an F1- score of 1.0 on holdout data (50 eyes total, 10 FTMH+), compared ImageNet pre-trained models, respectively. These results demonstrate that even limited data may be applied toward self-supervised pre- training to substantially improve performance for FTMH classification, indicating applicability toward other OCT-based problems.","PeriodicalId":501390,"journal":{"name":"medRxiv - Ophthalmology","volume":"58 9","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"medRxiv - Ophthalmology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2023.11.14.23298513","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
There is a growing interest in using computer-assisted models for the detection of macular conditions using optical coherence tomography (OCT) data. As the quantity of clinical scan data of specific conditions is limited, these models are typically developed by fine-tuning a generalized network to classify specific macular conditions of interest. Full thickness macular holes (FTMH) present a condition requiring timely surgical intervention to prevent permanent vision loss. Other works on automated FTMH classification have tended to use supervised ImageNet pre-trained networks with good results but leave room for improvement. In this paper, we develop a model for FTMH classification using OCT slices around the central foveal region to pre-train a naïve network using contrastive self-supervised learning. We found that self-supervised pre-trained networks outperform ImageNet pre-trained networks despite a small training set size (284 eyes total, 51 FTMH+ eyes, 3 slices from each eye). 3D spatial contrast pre-training yields a model with an F1- score of 1.0 on holdout data (50 eyes total, 10 FTMH+), compared ImageNet pre-trained models, respectively. These results demonstrate that even limited data may be applied toward self-supervised pre- training to substantially improve performance for FTMH classification, indicating applicability toward other OCT-based problems.