{"title":"基于话务员在水平面上的空间分离的定位和语音识别的频谱权值。","authors":"Emily Buss, Richard Freyman","doi":"10.1121/10.0037072","DOIUrl":null,"url":null,"abstract":"<p><p>Some previous research has suggested that sound source localization may not rely on the same cues that support the segregation of speech produced by talkers separated in space. The present experiments evaluated spectral weights for the spatial cues underlying these two tasks by filtering stimuli into 1-octave-wide bands and dispersing them on the horizontal plane. Target stimuli were 100-ms bursts of speech-shaped noise or words produced by 24 male and female talkers, and maskers (when present) were sequences of words. For localization in quiet, weights differed depending on the midpoint and band dispersion range, but they were similar for speech and noise stimuli. For bands dispersed between -15° and +15°, weights peaked at 500 and 1000 Hz. Introducing a speech masker changed the magnitude of weights for localization, but not the relative weight by frequency. For speech-in-speech recognition, sequences of masker words produced predominantly informational masking, such that participants had to rely on spatial cues to segregate the target. As for localization, recognition appeared to rely predominantly on spatial cues in the 500- and 1000-Hz bands. Trial-by-trial data suggest that correct word recognition relied on differences in perceived location of target and masker speech for some but not for all participants.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"158 1","pages":"186-200"},"PeriodicalIF":2.1000,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Spectral weights for localization and speech-in-speech recognition with spatial separation of talkers on the horizontal plane.\",\"authors\":\"Emily Buss, Richard Freyman\",\"doi\":\"10.1121/10.0037072\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Some previous research has suggested that sound source localization may not rely on the same cues that support the segregation of speech produced by talkers separated in space. The present experiments evaluated spectral weights for the spatial cues underlying these two tasks by filtering stimuli into 1-octave-wide bands and dispersing them on the horizontal plane. Target stimuli were 100-ms bursts of speech-shaped noise or words produced by 24 male and female talkers, and maskers (when present) were sequences of words. For localization in quiet, weights differed depending on the midpoint and band dispersion range, but they were similar for speech and noise stimuli. For bands dispersed between -15° and +15°, weights peaked at 500 and 1000 Hz. Introducing a speech masker changed the magnitude of weights for localization, but not the relative weight by frequency. For speech-in-speech recognition, sequences of masker words produced predominantly informational masking, such that participants had to rely on spatial cues to segregate the target. As for localization, recognition appeared to rely predominantly on spatial cues in the 500- and 1000-Hz bands. Trial-by-trial data suggest that correct word recognition relied on differences in perceived location of target and masker speech for some but not for all participants.</p>\",\"PeriodicalId\":17168,\"journal\":{\"name\":\"Journal of the Acoustical Society of America\",\"volume\":\"158 1\",\"pages\":\"186-200\"},\"PeriodicalIF\":2.1000,\"publicationDate\":\"2025-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of the Acoustical Society of America\",\"FirstCategoryId\":\"101\",\"ListUrlMain\":\"https://doi.org/10.1121/10.0037072\",\"RegionNum\":2,\"RegionCategory\":\"物理与天体物理\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ACOUSTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Acoustical Society of America","FirstCategoryId":"101","ListUrlMain":"https://doi.org/10.1121/10.0037072","RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ACOUSTICS","Score":null,"Total":0}
Spectral weights for localization and speech-in-speech recognition with spatial separation of talkers on the horizontal plane.
Some previous research has suggested that sound source localization may not rely on the same cues that support the segregation of speech produced by talkers separated in space. The present experiments evaluated spectral weights for the spatial cues underlying these two tasks by filtering stimuli into 1-octave-wide bands and dispersing them on the horizontal plane. Target stimuli were 100-ms bursts of speech-shaped noise or words produced by 24 male and female talkers, and maskers (when present) were sequences of words. For localization in quiet, weights differed depending on the midpoint and band dispersion range, but they were similar for speech and noise stimuli. For bands dispersed between -15° and +15°, weights peaked at 500 and 1000 Hz. Introducing a speech masker changed the magnitude of weights for localization, but not the relative weight by frequency. For speech-in-speech recognition, sequences of masker words produced predominantly informational masking, such that participants had to rely on spatial cues to segregate the target. As for localization, recognition appeared to rely predominantly on spatial cues in the 500- and 1000-Hz bands. Trial-by-trial data suggest that correct word recognition relied on differences in perceived location of target and masker speech for some but not for all participants.
期刊介绍:
Since 1929 The Journal of the Acoustical Society of America has been the leading source of theoretical and experimental research results in the broad interdisciplinary study of sound. Subject coverage includes: linear and nonlinear acoustics; aeroacoustics, underwater sound and acoustical oceanography; ultrasonics and quantum acoustics; architectural and structural acoustics and vibration; speech, music and noise; psychology and physiology of hearing; engineering acoustics, transduction; bioacoustics, animal bioacoustics.