Maria Villamil, Allie C Schneider, Jiahe Cui, Laura K Young, Hannah E Smithson
{"title":"贡献演讲一:利用计算机视觉从光感受器马赛克的AOSLO图像中检测和表征微眼跳。","authors":"Maria Villamil, Allie C Schneider, Jiahe Cui, Laura K Young, Hannah E Smithson","doi":"10.1167/jov.25.5.5","DOIUrl":null,"url":null,"abstract":"<p><p>Fixational eye movements (FEMs), especially microsaccades (MS), are promising biomarkers of neurodegenerative disease. In vivo images of the photoreceptor mosaic acquired using an Adaptive Optics Scanning Laser Ophthalmoscope (AOSLO) are systematically distorted by eye motion. Most methods to extract FEMs from AOSLO data rely on comparison to a motion-free reference, giving eye-position as a function of time. MS are subsequently identified using adaptive velocity thresholds (Engbert & Kliegl, 2003). We use computer vision and machine learning (ML) for detection and characterisation of MS directly from raw AOSLO images. For training and validation, we use Emulated Retinal Image CApture (ERICA), an open-source tool to generate synthetic AOSLO datasets of retinal images and ground-truth velocity profiles (Young & Smithson, 2021). To classify regions of AOSLO images that contain a MS, images were divided into a grid of 32-by-32-pixel sub-images. Predictions from rows of sub-images aligned with the fast-scan of the AOSLO were combined, giving 1ms resolution. Model performance was high (F1 scores >0.92) across plausible MS displacement magnitudes and angles, with most errors close to the velocity threshold for classification. Direct velocity predictions were also derived from regression ML models. We show that ML models can be systematically adapted for generalisation to real in vivo images, allowing characterisation of MS at much finer spatial scales than video-based eye-trackers.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 5","pages":"5"},"PeriodicalIF":2.3000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Contributed Talks I: Detecting and characterising microsaccades from AOSLO images of the photoreceptor mosaic using computer vision.\",\"authors\":\"Maria Villamil, Allie C Schneider, Jiahe Cui, Laura K Young, Hannah E Smithson\",\"doi\":\"10.1167/jov.25.5.5\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Fixational eye movements (FEMs), especially microsaccades (MS), are promising biomarkers of neurodegenerative disease. In vivo images of the photoreceptor mosaic acquired using an Adaptive Optics Scanning Laser Ophthalmoscope (AOSLO) are systematically distorted by eye motion. Most methods to extract FEMs from AOSLO data rely on comparison to a motion-free reference, giving eye-position as a function of time. MS are subsequently identified using adaptive velocity thresholds (Engbert & Kliegl, 2003). We use computer vision and machine learning (ML) for detection and characterisation of MS directly from raw AOSLO images. For training and validation, we use Emulated Retinal Image CApture (ERICA), an open-source tool to generate synthetic AOSLO datasets of retinal images and ground-truth velocity profiles (Young & Smithson, 2021). To classify regions of AOSLO images that contain a MS, images were divided into a grid of 32-by-32-pixel sub-images. Predictions from rows of sub-images aligned with the fast-scan of the AOSLO were combined, giving 1ms resolution. Model performance was high (F1 scores >0.92) across plausible MS displacement magnitudes and angles, with most errors close to the velocity threshold for classification. Direct velocity predictions were also derived from regression ML models. We show that ML models can be systematically adapted for generalisation to real in vivo images, allowing characterisation of MS at much finer spatial scales than video-based eye-trackers.</p>\",\"PeriodicalId\":49955,\"journal\":{\"name\":\"Journal of Vision\",\"volume\":\"25 5\",\"pages\":\"5\"},\"PeriodicalIF\":2.3000,\"publicationDate\":\"2025-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Vision\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1167/jov.25.5.5\",\"RegionNum\":4,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"OPHTHALMOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Vision","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1167/jov.25.5.5","RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
Contributed Talks I: Detecting and characterising microsaccades from AOSLO images of the photoreceptor mosaic using computer vision.
Fixational eye movements (FEMs), especially microsaccades (MS), are promising biomarkers of neurodegenerative disease. In vivo images of the photoreceptor mosaic acquired using an Adaptive Optics Scanning Laser Ophthalmoscope (AOSLO) are systematically distorted by eye motion. Most methods to extract FEMs from AOSLO data rely on comparison to a motion-free reference, giving eye-position as a function of time. MS are subsequently identified using adaptive velocity thresholds (Engbert & Kliegl, 2003). We use computer vision and machine learning (ML) for detection and characterisation of MS directly from raw AOSLO images. For training and validation, we use Emulated Retinal Image CApture (ERICA), an open-source tool to generate synthetic AOSLO datasets of retinal images and ground-truth velocity profiles (Young & Smithson, 2021). To classify regions of AOSLO images that contain a MS, images were divided into a grid of 32-by-32-pixel sub-images. Predictions from rows of sub-images aligned with the fast-scan of the AOSLO were combined, giving 1ms resolution. Model performance was high (F1 scores >0.92) across plausible MS displacement magnitudes and angles, with most errors close to the velocity threshold for classification. Direct velocity predictions were also derived from regression ML models. We show that ML models can be systematically adapted for generalisation to real in vivo images, allowing characterisation of MS at much finer spatial scales than video-based eye-trackers.
期刊介绍:
Exploring all aspects of biological visual function, including spatial vision, perception,
low vision, color vision and more, spanning the fields of neuroscience, psychology and psychophysics.