Hyun-Tae Choi, Eidmann Ammienn Bin Eh Mi, Bum-Kyu Kim, Won-Du Chang
{"title":"Electrooculography signal generation with conditional diffusion models for eye movement classification","authors":"Hyun-Tae Choi, Eidmann Ammienn Bin Eh Mi, Bum-Kyu Kim, Won-Du Chang","doi":"10.1016/j.bspc.2025.108211","DOIUrl":null,"url":null,"abstract":"<div><div>Electrooculography (EOG) is a biosignal that encodes the directional information of eye movements and is widely used in eye-tracking applications and human-computer interaction. These applications provide intuitive, accessible interfaces, making EOG valuable as a communication aid. Among these applications, eye writing—an approach in which users draw characters using eye movements—is distinguished for its ability to convey significantly more information than traditional methods. This technique has potential applications for individuals who rely on eye movements for communication, including those with amyotrophic lateral sclerosis. However, achieving high accuracy in eye writing typically requires deep learning models constrained by ethical and legal challenges in collecting large EOG datasets. In this study, we developed a diffusion-model-based generative framework for eye-written character recognition under limited-data conditions, addressing challenges in both accuracy and data availability. The model was implemented using conditional vectorized class information to further increase the diversity of the generated data by indicating the class. The effectiveness of the proposed method was assessed through visual representations of generated signals and comparison of classification accuracies. The model outperformed the generative adversarial network in terms of the visual quality of generated data. It also achieved classification accuracies of 94.63% for Arabic numerals and 80.36% for Japanese Katakana strokes when trained with ninefold data at intermediate steps. Despite having significantly more parameters, it achieved shorter inference time than TimeGAN, further demonstrating computational efficiency and feasibility. Additionally, adapting the proposed method to other bioelectric signals demonstrates significant potential, suggesting a flexible framework for future biosignal research and applications.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"110 ","pages":"Article 108211"},"PeriodicalIF":4.9000,"publicationDate":"2025-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical Signal Processing and Control","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1746809425007220","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Electrooculography (EOG) is a biosignal that encodes the directional information of eye movements and is widely used in eye-tracking applications and human-computer interaction. These applications provide intuitive, accessible interfaces, making EOG valuable as a communication aid. Among these applications, eye writing—an approach in which users draw characters using eye movements—is distinguished for its ability to convey significantly more information than traditional methods. This technique has potential applications for individuals who rely on eye movements for communication, including those with amyotrophic lateral sclerosis. However, achieving high accuracy in eye writing typically requires deep learning models constrained by ethical and legal challenges in collecting large EOG datasets. In this study, we developed a diffusion-model-based generative framework for eye-written character recognition under limited-data conditions, addressing challenges in both accuracy and data availability. The model was implemented using conditional vectorized class information to further increase the diversity of the generated data by indicating the class. The effectiveness of the proposed method was assessed through visual representations of generated signals and comparison of classification accuracies. The model outperformed the generative adversarial network in terms of the visual quality of generated data. It also achieved classification accuracies of 94.63% for Arabic numerals and 80.36% for Japanese Katakana strokes when trained with ninefold data at intermediate steps. Despite having significantly more parameters, it achieved shorter inference time than TimeGAN, further demonstrating computational efficiency and feasibility. Additionally, adapting the proposed method to other bioelectric signals demonstrates significant potential, suggesting a flexible framework for future biosignal research and applications.
期刊介绍:
Biomedical Signal Processing and Control aims to provide a cross-disciplinary international forum for the interchange of information on research in the measurement and analysis of signals and images in clinical medicine and the biological sciences. Emphasis is placed on contributions dealing with the practical, applications-led research on the use of methods and devices in clinical diagnosis, patient monitoring and management.
Biomedical Signal Processing and Control reflects the main areas in which these methods are being used and developed at the interface of both engineering and clinical science. The scope of the journal is defined to include relevant review papers, technical notes, short communications and letters. Tutorial papers and special issues will also be published.