Mikhail Kulyabin, Aleksei Zhdanov, Irene O Lee, David H Skuse, Dorothy A Thompson, Andreas Maier, Paul A Constable
{"title":"Synthetic electroretinogram signal generation using a conditional generative adversarial network.","authors":"Mikhail Kulyabin, Aleksei Zhdanov, Irene O Lee, David H Skuse, Dorothy A Thompson, Andreas Maier, Paul A Constable","doi":"10.1007/s10633-025-10019-0","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>The electroretinogram (ERG) records the functional response of the retina. In some neurological conditions, the ERG waveform may be altered and could support biomarker discovery. In heterogeneous or rare populations, where either large data sets or the availability of data may be a challenge, synthetic signals with Artificial Intelligence (AI) may help to mitigate against these factors to support classification models.</p><p><strong>Methods: </strong>This approach was tested using a publicly available dataset of real ERGs, n = 560 (ASD) and n = 498 (Control) recorded at 9 different flash strengths from n = 18 ASD (mean age 12.2 ± 2.7 years) and n = 31 Controls (mean age 11.8 ± 3.3 years) that were augmented with synthetic waveforms, generated through a Conditional Generative Adversarial Network. Two deep learning models were used to classify the groups using either the real only or combined real and synthetic ERGs. One was a Time Series Transformer (with waveforms in their original form) and the second was a Visual Transformer model utilizing images of the wavelets derived from a Continuous Wavelet Transform of the ERGs. Model performance at classifying the groups was evaluated with Balanced Accuracy (BA) as the main outcome measure.</p><p><strong>Results: </strong>The BA improved from 0.756 to 0.879 when synthetic ERGs were included across all recordings for the training of the Time Series Transformer. This model also achieved the best performance with a BA of 0.89 using real and synthetic waveforms from a single flash strength of 0.95 log cd s m<sup>-2</sup>.</p><p><strong>Conclusions: </strong>The improved performance of the deep learning models with synthetic waveforms supports the application of AI to improve group classification with ERG recordings.</p>","PeriodicalId":11207,"journal":{"name":"Documenta Ophthalmologica","volume":" ","pages":""},"PeriodicalIF":2.6000,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Documenta Ophthalmologica","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s10633-025-10019-0","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: The electroretinogram (ERG) records the functional response of the retina. In some neurological conditions, the ERG waveform may be altered and could support biomarker discovery. In heterogeneous or rare populations, where either large data sets or the availability of data may be a challenge, synthetic signals with Artificial Intelligence (AI) may help to mitigate against these factors to support classification models.
Methods: This approach was tested using a publicly available dataset of real ERGs, n = 560 (ASD) and n = 498 (Control) recorded at 9 different flash strengths from n = 18 ASD (mean age 12.2 ± 2.7 years) and n = 31 Controls (mean age 11.8 ± 3.3 years) that were augmented with synthetic waveforms, generated through a Conditional Generative Adversarial Network. Two deep learning models were used to classify the groups using either the real only or combined real and synthetic ERGs. One was a Time Series Transformer (with waveforms in their original form) and the second was a Visual Transformer model utilizing images of the wavelets derived from a Continuous Wavelet Transform of the ERGs. Model performance at classifying the groups was evaluated with Balanced Accuracy (BA) as the main outcome measure.
Results: The BA improved from 0.756 to 0.879 when synthetic ERGs were included across all recordings for the training of the Time Series Transformer. This model also achieved the best performance with a BA of 0.89 using real and synthetic waveforms from a single flash strength of 0.95 log cd s m-2.
Conclusions: The improved performance of the deep learning models with synthetic waveforms supports the application of AI to improve group classification with ERG recordings.
期刊介绍:
Documenta Ophthalmologica is an official publication of the International Society for Clinical Electrophysiology of Vision. The purpose of the journal is to promote the understanding and application of clinical electrophysiology of vision. Documenta Ophthalmologica will publish reviews, research articles, technical notes, brief reports and case studies which inform the readers about basic and clinical sciences related to visual electrodiagnosis and means to improve diagnosis and clinical management of patients using visual electrophysiology. Studies may involve animals or humans. In either case appropriate care must be taken to follow the Declaration of Helsinki for human subject or appropriate humane standards of animal care (e.g., the ARVO standards on Animal Care and Use).