Denise Alonso-Vázquez, Omar Mendoza-Montoya, Ricardo Caraza, Hector R Martinez, Javier M Antelis
{"title":"从发音到想象:用多条件脑电数据改进语音解码。","authors":"Denise Alonso-Vázquez, Omar Mendoza-Montoya, Ricardo Caraza, Hector R Martinez, Javier M Antelis","doi":"10.3389/fninf.2025.1583428","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong><i>Imagined speech</i> decoding using EEG holds promising applications for individuals with motor neuron diseases, although its performance remains limited due to small dataset sizes and the absence of sensory feedback. Here, we investigated whether incorporating EEG data from <i>overt</i> (pronounced) speech could enhance <i>imagined speech</i> classification.</p><p><strong>Methods: </strong>Our approach systematically compares four classification scenarios by modifying the training dataset: intra-subject (using only <i>imagined speech</i>, combining <i>overt</i> and <i>imagined speech</i>, and using only <i>overt speech</i>) and multi-subject (combining <i>overt speech</i> data from different participants with the <i>imagined speech</i> of the target participant). We implemented all scenarios using the convolutional neural network EEGNet. To this end, twenty-four healthy participants pronounced and imagined five Spanish words.</p><p><strong>Results: </strong>In binary word-pair classifications, combining <i>overt</i> and <i>imagined speech</i> data in the intra-subject scenario led to accuracy improvements of 3%-5.17% in four out of 10 word pairs, compared to training with <i>imagined speech</i> only. Although the highest individual accuracy (95%) was achieved with <i>imagined speech</i> alone, the inclusion of <i>overt speech</i> data allowed more participants to surpass 70% accuracy, increasing from 10 (<i>imagined only</i>) to 15 participants. In the intra-subject multi-class scenario, combining <i>overt</i> and <i>imagined speech</i> did not yield statistically significant improvements over using <i>imagined speech</i> exclusively.</p><p><strong>Discussion: </strong>Finally, we observed that features such as word length, phonological complexity, and frequency of use contributed to higher discriminability between certain <i>imagined</i> word pairs. These findings suggest that incorporating <i>overt speech</i> data can improve <i>imagined speech</i> decoding in individualized models, offering a feasible strategy to support the early adoption of brain-computer interfaces before speech deterioration occurs in individuals with motor neuron diseases.</p>","PeriodicalId":12462,"journal":{"name":"Frontiers in Neuroinformatics","volume":"19 ","pages":"1583428"},"PeriodicalIF":2.5000,"publicationDate":"2025-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12245923/pdf/","citationCount":"0","resultStr":"{\"title\":\"From pronounced to imagined: improving speech decoding with multi-condition EEG data.\",\"authors\":\"Denise Alonso-Vázquez, Omar Mendoza-Montoya, Ricardo Caraza, Hector R Martinez, Javier M Antelis\",\"doi\":\"10.3389/fninf.2025.1583428\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Introduction: </strong><i>Imagined speech</i> decoding using EEG holds promising applications for individuals with motor neuron diseases, although its performance remains limited due to small dataset sizes and the absence of sensory feedback. Here, we investigated whether incorporating EEG data from <i>overt</i> (pronounced) speech could enhance <i>imagined speech</i> classification.</p><p><strong>Methods: </strong>Our approach systematically compares four classification scenarios by modifying the training dataset: intra-subject (using only <i>imagined speech</i>, combining <i>overt</i> and <i>imagined speech</i>, and using only <i>overt speech</i>) and multi-subject (combining <i>overt speech</i> data from different participants with the <i>imagined speech</i> of the target participant). We implemented all scenarios using the convolutional neural network EEGNet. To this end, twenty-four healthy participants pronounced and imagined five Spanish words.</p><p><strong>Results: </strong>In binary word-pair classifications, combining <i>overt</i> and <i>imagined speech</i> data in the intra-subject scenario led to accuracy improvements of 3%-5.17% in four out of 10 word pairs, compared to training with <i>imagined speech</i> only. Although the highest individual accuracy (95%) was achieved with <i>imagined speech</i> alone, the inclusion of <i>overt speech</i> data allowed more participants to surpass 70% accuracy, increasing from 10 (<i>imagined only</i>) to 15 participants. In the intra-subject multi-class scenario, combining <i>overt</i> and <i>imagined speech</i> did not yield statistically significant improvements over using <i>imagined speech</i> exclusively.</p><p><strong>Discussion: </strong>Finally, we observed that features such as word length, phonological complexity, and frequency of use contributed to higher discriminability between certain <i>imagined</i> word pairs. These findings suggest that incorporating <i>overt speech</i> data can improve <i>imagined speech</i> decoding in individualized models, offering a feasible strategy to support the early adoption of brain-computer interfaces before speech deterioration occurs in individuals with motor neuron diseases.</p>\",\"PeriodicalId\":12462,\"journal\":{\"name\":\"Frontiers in Neuroinformatics\",\"volume\":\"19 \",\"pages\":\"1583428\"},\"PeriodicalIF\":2.5000,\"publicationDate\":\"2025-06-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12245923/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Frontiers in Neuroinformatics\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.3389/fninf.2025.1583428\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q2\",\"JCRName\":\"MATHEMATICAL & COMPUTATIONAL BIOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Neuroinformatics","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.3389/fninf.2025.1583428","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"MATHEMATICAL & COMPUTATIONAL BIOLOGY","Score":null,"Total":0}
From pronounced to imagined: improving speech decoding with multi-condition EEG data.
Introduction: Imagined speech decoding using EEG holds promising applications for individuals with motor neuron diseases, although its performance remains limited due to small dataset sizes and the absence of sensory feedback. Here, we investigated whether incorporating EEG data from overt (pronounced) speech could enhance imagined speech classification.
Methods: Our approach systematically compares four classification scenarios by modifying the training dataset: intra-subject (using only imagined speech, combining overt and imagined speech, and using only overt speech) and multi-subject (combining overt speech data from different participants with the imagined speech of the target participant). We implemented all scenarios using the convolutional neural network EEGNet. To this end, twenty-four healthy participants pronounced and imagined five Spanish words.
Results: In binary word-pair classifications, combining overt and imagined speech data in the intra-subject scenario led to accuracy improvements of 3%-5.17% in four out of 10 word pairs, compared to training with imagined speech only. Although the highest individual accuracy (95%) was achieved with imagined speech alone, the inclusion of overt speech data allowed more participants to surpass 70% accuracy, increasing from 10 (imagined only) to 15 participants. In the intra-subject multi-class scenario, combining overt and imagined speech did not yield statistically significant improvements over using imagined speech exclusively.
Discussion: Finally, we observed that features such as word length, phonological complexity, and frequency of use contributed to higher discriminability between certain imagined word pairs. These findings suggest that incorporating overt speech data can improve imagined speech decoding in individualized models, offering a feasible strategy to support the early adoption of brain-computer interfaces before speech deterioration occurs in individuals with motor neuron diseases.
期刊介绍:
Frontiers in Neuroinformatics publishes rigorously peer-reviewed research on the development and implementation of numerical/computational models and analytical tools used to share, integrate and analyze experimental data and advance theories of the nervous system functions. Specialty Chief Editors Jan G. Bjaalie at the University of Oslo and Sean L. Hill at the École Polytechnique Fédérale de Lausanne are supported by an outstanding Editorial Board of international experts. This multidisciplinary open-access journal is at the forefront of disseminating and communicating scientific knowledge and impactful discoveries to researchers, academics and the public worldwide.
Neuroscience is being propelled into the information age as the volume of information explodes, demanding organization and synthesis. Novel synthesis approaches are opening up a new dimension for the exploration of the components of brain elements and systems and the vast number of variables that underlie their functions. Neural data is highly heterogeneous with complex inter-relations across multiple levels, driving the need for innovative organizing and synthesizing approaches from genes to cognition, and covering a range of species and disease states.
Frontiers in Neuroinformatics therefore welcomes submissions on existing neuroscience databases, development of data and knowledge bases for all levels of neuroscience, applications and technologies that can facilitate data sharing (interoperability, formats, terminologies, and ontologies), and novel tools for data acquisition, analyses, visualization, and dissemination of nervous system data. Our journal welcomes submissions on new tools (software and hardware) that support brain modeling, and the merging of neuroscience databases with brain models used for simulation and visualization.