Aurélie de Borman, Benjamin Wittevrongel, Bob Van Dyck, Kato Van Rooy, Evelien Carrette, Alfred Meurs, Dirk Van Roost, Marc M Van Hulle
{"title":"Speech mode classification from electrocorticography: transfer between electrodes and participants.","authors":"Aurélie de Borman, Benjamin Wittevrongel, Bob Van Dyck, Kato Van Rooy, Evelien Carrette, Alfred Meurs, Dirk Van Roost, Marc M Van Hulle","doi":"10.1088/1741-2552/adf2de","DOIUrl":null,"url":null,"abstract":"<p><p><i>Objective.</i>Speech brain-computer interfaces (BCIs) aim to restore communication for individuals who have lost the ability to speak by interpreting their brain activity and decoding the intended speech. As an initial component of these decoders, speech detectors have been developed to distinguish between the intent to speak and silence. However, it is important that these detectors account for real-life scenarios in which users may engage language-related brain areas-such as during reading or listening-without any intention to speak.<i>Approach.</i>In this study, we analyze the interplay between different speech modes: speaking, listening, imagining speaking, reading and mouthing. We gathered a large dataset of 29 participants implanted with electrocorticography electrodes and developed a speech mode classifier. We also assessed how well classifiers trained on data from a specific participant transfer to other participants, both in the case of a single- and multi-electrode classifier.<i>Main results.</i>High accuracy was achieved using linear classifiers, for both single-electrode and multi-electrode configurations. Single-electrode classification reached 88.89% accuracy and multi-electrode classification 96.49% accuracy in distinguishing among three classes (speaking, listening, and silence). The best performing electrodes were located on the superior temporal gyrus and sensorimotor cortex. We found that single-electrode classifiers could be transferred across recording sites. For multi-electrode classifiers, we observed that transfer performance was higher for binary classifiers compared to multiclass classifiers, with the optimal source subject of the binary classifiers depending on the speech modes being classified.<i>Significance</i>Accurately detecting speech from brain signals is essential to prevent spurious outputs from a speech BCI and to advance its use beyond lab settings. To achieve this objective, the transfer between participants is particularly valuable as it can reduce training time, especially in cases where subject training is challenging.</p>","PeriodicalId":94096,"journal":{"name":"Journal of neural engineering","volume":" ","pages":""},"PeriodicalIF":3.8000,"publicationDate":"2025-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of neural engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/1741-2552/adf2de","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Objective.Speech brain-computer interfaces (BCIs) aim to restore communication for individuals who have lost the ability to speak by interpreting their brain activity and decoding the intended speech. As an initial component of these decoders, speech detectors have been developed to distinguish between the intent to speak and silence. However, it is important that these detectors account for real-life scenarios in which users may engage language-related brain areas-such as during reading or listening-without any intention to speak.Approach.In this study, we analyze the interplay between different speech modes: speaking, listening, imagining speaking, reading and mouthing. We gathered a large dataset of 29 participants implanted with electrocorticography electrodes and developed a speech mode classifier. We also assessed how well classifiers trained on data from a specific participant transfer to other participants, both in the case of a single- and multi-electrode classifier.Main results.High accuracy was achieved using linear classifiers, for both single-electrode and multi-electrode configurations. Single-electrode classification reached 88.89% accuracy and multi-electrode classification 96.49% accuracy in distinguishing among three classes (speaking, listening, and silence). The best performing electrodes were located on the superior temporal gyrus and sensorimotor cortex. We found that single-electrode classifiers could be transferred across recording sites. For multi-electrode classifiers, we observed that transfer performance was higher for binary classifiers compared to multiclass classifiers, with the optimal source subject of the binary classifiers depending on the speech modes being classified.SignificanceAccurately detecting speech from brain signals is essential to prevent spurious outputs from a speech BCI and to advance its use beyond lab settings. To achieve this objective, the transfer between participants is particularly valuable as it can reduce training time, especially in cases where subject training is challenging.