枕颞皮层及其他部位听觉和视觉言语的语音表征。

IF 4.4 2区 医学 Q1 NEUROSCIENCES
Alice Van Audenhaege, Stefania Mattioni, Filippo Cerpelloni, Remi Gau, Arnaud Szmalec, Olivier Collignon
{"title":"枕颞皮层及其他部位听觉和视觉言语的语音表征。","authors":"Alice Van Audenhaege, Stefania Mattioni, Filippo Cerpelloni, Remi Gau, Arnaud Szmalec, Olivier Collignon","doi":"10.1523/JNEUROSCI.1415-24.2025","DOIUrl":null,"url":null,"abstract":"<p><p>Speech is a multisensory signal that can be extracted from the voice and the lips. Previous studies suggested that occipital and temporal regions encode both auditory and visual speech features but their location and nature remain unclear. We characterized brain activity using fMRI (13 males and 11 females) to functionally and individually define bilateral fusiform face areas (FFA), the left word-selective ventral occipito-temporal cortex (word-VOTC), an audiovisual speech region in the left superior temporal sulcus (lSTS); and control regions in bilateral scene-selective parahippocampal place areas (PPA). In these regions, we performed multivariate pattern classification of corresponding phonemes (speech sounds) and visemes (lip movements). We observed that the word-VOTC and lSTS represent phonological information from both vision and sounds. The multisensory nature of phonological representations appeared selective to the word-VOTC, as we found viseme but not phoneme representation in adjacent FFA, while PPA did not encode phonology in any modality. Interestingly, cross-modal decoding revealed aligned phonological representations across the senses in lSTS, but not in word-VOTC. A whole-brain cross-modal searchlight analysis additionally revealed aligned audiovisual phonological representations in bilateral pSTS and left somato-motor cortex overlapping with oro-facial articulators. Altogether, our results demonstrate that auditory and visual phonology are represented in the word-VOTC, extending its functional coding beyond orthography. The geometries of auditory and visual representations do not align in the word-VOTC as they do in the STS and left somato-motor cortex, suggesting distinct representations across a distributed multisensory phonological network.</p>","PeriodicalId":50114,"journal":{"name":"Journal of Neuroscience","volume":" ","pages":""},"PeriodicalIF":4.4000,"publicationDate":"2025-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12199548/pdf/","citationCount":"0","resultStr":"{\"title\":\"Phonological Representations of Auditory and Visual Speech in the Occipito-temporal Cortex and Beyond.\",\"authors\":\"Alice Van Audenhaege, Stefania Mattioni, Filippo Cerpelloni, Remi Gau, Arnaud Szmalec, Olivier Collignon\",\"doi\":\"10.1523/JNEUROSCI.1415-24.2025\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Speech is a multisensory signal that can be extracted from the voice and the lips. Previous studies suggested that occipital and temporal regions encode both auditory and visual speech features but their location and nature remain unclear. We characterized brain activity using fMRI (13 males and 11 females) to functionally and individually define bilateral fusiform face areas (FFA), the left word-selective ventral occipito-temporal cortex (word-VOTC), an audiovisual speech region in the left superior temporal sulcus (lSTS); and control regions in bilateral scene-selective parahippocampal place areas (PPA). In these regions, we performed multivariate pattern classification of corresponding phonemes (speech sounds) and visemes (lip movements). We observed that the word-VOTC and lSTS represent phonological information from both vision and sounds. The multisensory nature of phonological representations appeared selective to the word-VOTC, as we found viseme but not phoneme representation in adjacent FFA, while PPA did not encode phonology in any modality. Interestingly, cross-modal decoding revealed aligned phonological representations across the senses in lSTS, but not in word-VOTC. A whole-brain cross-modal searchlight analysis additionally revealed aligned audiovisual phonological representations in bilateral pSTS and left somato-motor cortex overlapping with oro-facial articulators. Altogether, our results demonstrate that auditory and visual phonology are represented in the word-VOTC, extending its functional coding beyond orthography. The geometries of auditory and visual representations do not align in the word-VOTC as they do in the STS and left somato-motor cortex, suggesting distinct representations across a distributed multisensory phonological network.</p>\",\"PeriodicalId\":50114,\"journal\":{\"name\":\"Journal of Neuroscience\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":4.4000,\"publicationDate\":\"2025-06-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12199548/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Neuroscience\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1523/JNEUROSCI.1415-24.2025\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"NEUROSCIENCES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Neuroscience","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1523/JNEUROSCI.1415-24.2025","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0

摘要

语音是一种多感官信号,可以从声音和嘴唇中提取。先前的研究表明,枕叶和颞叶区域编码听觉和视觉语言特征,但它们的位置和性质尚不清楚。我们使用功能磁共振成像(fMRI)来表征大脑活动(男性和女性),以功能和单独定义双侧梭状脸区(FFA),左侧词选择性腹侧枕颞皮质(word-VOTC),左侧颞上沟(lSTS)的视听语言区域和双侧情景选择性海马旁地方区(PPA)的控制区域。在这些区域,我们对相应的音素(语音)和音素(唇部运动)进行了多元模式分类。我们观察到,单词- votc和lSTS代表了视觉和声音的语音信息。语音表征的多感官特性似乎对单词- votc有选择性,因为我们在相邻的FFA中发现了音素表征而不是音素表征,而PPA在任何情态中都没有编码音素。有趣的是,跨模态解码在lSTS中发现了跨感官的一致语音表征,而在word-VOTC中则没有。全脑跨模态探照灯分析还显示,双侧pSTS和左侧躯体运动皮层的视听语音表征与口面部发音重叠。总之,我们的研究结果表明,听觉和视觉音系在单词- votc中得到表征,扩展了其功能编码,超出了正字法。听觉和视觉表征的几何形状在单词- votc中并不像在STS和左侧躯体运动皮层中那样对齐,这表明在分布式语音网络中存在明显的多感觉表征。语音是一种多感官信号,可以从声音和嘴唇中提取。哪些大脑区域同时编码视觉和听觉语言表征?我们发现,书面单词选择腹侧枕颞(word-VOTC)和左侧颞上沟(lSTS)都处理来自语音和唇部运动的语音信息。然而,尽管lSTS将这些表征跨感官对齐,但单词- votc却没有,这表明不同的编码机制。这些发现将单词- votc的功能作用扩展到阅读之外。另一种全脑方法揭示了双侧颞上皮层和左侧躯体运动皮层的共同表征,表明了多感觉音系的分布式网络。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Phonological Representations of Auditory and Visual Speech in the Occipito-temporal Cortex and Beyond.

Speech is a multisensory signal that can be extracted from the voice and the lips. Previous studies suggested that occipital and temporal regions encode both auditory and visual speech features but their location and nature remain unclear. We characterized brain activity using fMRI (13 males and 11 females) to functionally and individually define bilateral fusiform face areas (FFA), the left word-selective ventral occipito-temporal cortex (word-VOTC), an audiovisual speech region in the left superior temporal sulcus (lSTS); and control regions in bilateral scene-selective parahippocampal place areas (PPA). In these regions, we performed multivariate pattern classification of corresponding phonemes (speech sounds) and visemes (lip movements). We observed that the word-VOTC and lSTS represent phonological information from both vision and sounds. The multisensory nature of phonological representations appeared selective to the word-VOTC, as we found viseme but not phoneme representation in adjacent FFA, while PPA did not encode phonology in any modality. Interestingly, cross-modal decoding revealed aligned phonological representations across the senses in lSTS, but not in word-VOTC. A whole-brain cross-modal searchlight analysis additionally revealed aligned audiovisual phonological representations in bilateral pSTS and left somato-motor cortex overlapping with oro-facial articulators. Altogether, our results demonstrate that auditory and visual phonology are represented in the word-VOTC, extending its functional coding beyond orthography. The geometries of auditory and visual representations do not align in the word-VOTC as they do in the STS and left somato-motor cortex, suggesting distinct representations across a distributed multisensory phonological network.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Journal of Neuroscience
Journal of Neuroscience 医学-神经科学
CiteScore
9.30
自引率
3.80%
发文量
1164
审稿时长
12 months
期刊介绍: JNeurosci (ISSN 0270-6474) is an official journal of the Society for Neuroscience. It is published weekly by the Society, fifty weeks a year, one volume a year. JNeurosci publishes papers on a broad range of topics of general interest to those working on the nervous system. Authors now have an Open Choice option for their published articles
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信