M. Wairagkar, L. Hochberg, D. Brandman, S. Stavisky
{"title":"解码背侧运动皮层皮层内神经活动的语音合成","authors":"M. Wairagkar, L. Hochberg, D. Brandman, S. Stavisky","doi":"10.1109/NER52421.2023.10123880","DOIUrl":null,"url":null,"abstract":"Losing the ability to speak due to brain injury or neurodegenerative diseases such as ALS can be debilitating. Brain-computer interfaces could potentially provide affected individuals a fast and intuitive way to communicate by decoding speech-related neural activity into a computer-synthesized voice. Current intracortical BCIs for communication using handwriting or point-and-click typing are substantially slower than natural speech and do not capture the full expressive range of speech. Recent studies have identified speech features from ECoG and sEEG recordings; however, intelligible speech synthesis has not yet been demonstrated. Our previous work has shown speech-related patterns in intracortical recordings from dorsal (arm/hand) motor cortex that enabled discrete word/phoneme classification. This motivates exploring an intracortical approach for continuous voice synthesis. Here, we present a neural decoding framework to synthesize speech by directly translating neural activity recorded from human motor cortex using intracortical multielectrode arrays into a low-dimensional speech feature space from which voice is synthesized.","PeriodicalId":201841,"journal":{"name":"2023 11th International IEEE/EMBS Conference on Neural Engineering (NER)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Synthesizing Speech by Decoding Intracortical Neural Activity from Dorsal Motor Cortex\",\"authors\":\"M. Wairagkar, L. Hochberg, D. Brandman, S. Stavisky\",\"doi\":\"10.1109/NER52421.2023.10123880\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Losing the ability to speak due to brain injury or neurodegenerative diseases such as ALS can be debilitating. Brain-computer interfaces could potentially provide affected individuals a fast and intuitive way to communicate by decoding speech-related neural activity into a computer-synthesized voice. Current intracortical BCIs for communication using handwriting or point-and-click typing are substantially slower than natural speech and do not capture the full expressive range of speech. Recent studies have identified speech features from ECoG and sEEG recordings; however, intelligible speech synthesis has not yet been demonstrated. Our previous work has shown speech-related patterns in intracortical recordings from dorsal (arm/hand) motor cortex that enabled discrete word/phoneme classification. This motivates exploring an intracortical approach for continuous voice synthesis. Here, we present a neural decoding framework to synthesize speech by directly translating neural activity recorded from human motor cortex using intracortical multielectrode arrays into a low-dimensional speech feature space from which voice is synthesized.\",\"PeriodicalId\":201841,\"journal\":{\"name\":\"2023 11th International IEEE/EMBS Conference on Neural Engineering (NER)\",\"volume\":\"23 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-04-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 11th International IEEE/EMBS Conference on Neural Engineering (NER)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NER52421.2023.10123880\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 11th International IEEE/EMBS Conference on Neural Engineering (NER)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NER52421.2023.10123880","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Synthesizing Speech by Decoding Intracortical Neural Activity from Dorsal Motor Cortex
Losing the ability to speak due to brain injury or neurodegenerative diseases such as ALS can be debilitating. Brain-computer interfaces could potentially provide affected individuals a fast and intuitive way to communicate by decoding speech-related neural activity into a computer-synthesized voice. Current intracortical BCIs for communication using handwriting or point-and-click typing are substantially slower than natural speech and do not capture the full expressive range of speech. Recent studies have identified speech features from ECoG and sEEG recordings; however, intelligible speech synthesis has not yet been demonstrated. Our previous work has shown speech-related patterns in intracortical recordings from dorsal (arm/hand) motor cortex that enabled discrete word/phoneme classification. This motivates exploring an intracortical approach for continuous voice synthesis. Here, we present a neural decoding framework to synthesize speech by directly translating neural activity recorded from human motor cortex using intracortical multielectrode arrays into a low-dimensional speech feature space from which voice is synthesized.