{"title":"Neural correlates of tonal loudness, intensity discrimination, and duration discrimination.","authors":"Shigeyuki Kuwada, Constantine Trahiotis","doi":"10.1121/10.0025874","DOIUrl":"10.1121/10.0025874","url":null,"abstract":"<p><p>A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Predicting underwater acoustic transmission loss in the SOFAR channel from ray trajectories via deep learning.","authors":"Haitao Wang, Shiwei Peng, Qunyi He, Xiangyang Zeng","doi":"10.1121/10.0025976","DOIUrl":"https://doi.org/10.1121/10.0025976","url":null,"abstract":"<p><p>Predicting acoustic transmission loss in the SOFAR channel faces challenges, such as excessively complex algorithms and computationally intensive calculations in classical methods. To address these challenges, a deep learning-based underwater acoustic transmission loss prediction method is proposed. By properly training a U-net-type convolutional neural network, the method can provide an accurate mapping between ray trajectories and the transmission loss over the problem domain. Verifications are performed in a SOFAR channel with Munk's sound speed profile. The results suggest that the method has potential to be used as a fast predicting model without sacrificing accuracy.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima
{"title":"Impact of speech rate on perception of vowel and consonant duration by bilinguals and monolinguals.","authors":"Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima","doi":"10.1121/10.0025862","DOIUrl":"10.1121/10.0025862","url":null,"abstract":"<p><p>The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya
{"title":"Reduced digit spans and ear dominance using dichotic digits in bimodal cochlear-implant users.","authors":"Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya","doi":"10.1121/10.0025977","DOIUrl":"10.1121/10.0025977","url":null,"abstract":"<p><p>Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11550484/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140899058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An open auscultation dataset for machine learning-based respiratory diagnosis studies.","authors":"Guanyu Zhou, Chengjian Liu, Xiaoguang Li, Sicong Liang, Ruichen Wang, Xun Huang","doi":"10.1121/10.0025851","DOIUrl":"10.1121/10.0025851","url":null,"abstract":"<p><p>Machine learning enabled auscultating diagnosis can provide promising solutions especially for prescreening purposes. The bottleneck for its potential success is that high-quality datasets for training are still scarce. An open auscultation dataset that consists of samples and annotations from patients and healthy individuals is established in this work for the respiratory diagnosis studies with machine learning, which is of both scientific importance and practical potential. A machine learning approach is examined to showcase the use of this new dataset for lung sound classifications with different diseases. The open dataset is available to the public online.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Clear speech effects in production of sentence-medial Mandarin lexical tonesa).","authors":"Jack Rittenberry, Irina A Shport","doi":"10.1121/10.0025991","DOIUrl":"10.1121/10.0025991","url":null,"abstract":"<p><p>Adding to limited research on clear speech in tone languages, productions of Mandarin lexical tones were examined in pentasyllabic sentences. Fourteen participants read sentences imagining a hard-of-hearing addressee or a friend in a casual social setting. Tones produced in clear speech had longer duration, higher intensity, and larger F0 values. This style effect was rarely modulated by tone, preceding tonal context, or syllable position, consistent with an overall signal enhancement strategy. Possible evidence for tone enhancement was observed only in one set of analysis for F0 minimum and F0 range, contrasting tones with low targets and tones with high targets.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141158923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Amara C Ezenwa, Matthew J Goupell, Sandra Gordon-Salant
{"title":"Cochlear-implant listeners benefit from training with time-compressed speech, even at advanced ages.","authors":"Amara C Ezenwa, Matthew J Goupell, Sandra Gordon-Salant","doi":"10.1121/10.0025431","DOIUrl":"10.1121/10.0025431","url":null,"abstract":"<p><p>This study evaluated whether adaptive training with time-compressed speech produces an age-dependent improvement in speech recognition in 14 adult cochlear-implant users. The protocol consisted of a pretest, 5 h of training, and a posttest using time-compressed speech and an adaptive procedure. There were significant improvements in time-compressed speech recognition at the posttest session following training (>5% in the average time-compressed speech recognition threshold) but no effects of age. These results are promising for the use of adaptive training in aural rehabilitation strategies for cochlear-implant users across the adult lifespan and possibly using speech signals, such as time-compressed speech, to train temporal processing.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11075136/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"English vowel recognition in multi-talker babbles mixed with different numbers of talkersa).","authors":"Xianhui Wang, Li Xu","doi":"10.1121/10.0025616","DOIUrl":"https://doi.org/10.1121/10.0025616","url":null,"abstract":"The present study examined English vowel recognition in multi-talker babbles (MTBs) in 20 normal-hearing, native-English-speaking adult listeners. Twelve vowels, embedded in the h-V-d structure, were presented in MTBs consisting of 1, 2, 4, 6, 8, 10, and 12 talkers (numbers of talkers [N]) and a speech-shaped noise at signal-to-noise ratios of -12, -6, and 0 dB. Results showed that vowel recognition performance was a non-monotonic function of N when signal-to-noise ratios were less favorable. The masking effects of MTBs on vowel recognition were most similar to consonant recognition but less so to word and sentence recognition reported in previous studies.","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140788246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust speed estimation for a moving harmonic acoustic source with a single stationary sensor.","authors":"Yixin Yang, Ningning Liang, Jianbo Zhou","doi":"10.1121/10.0025508","DOIUrl":"https://doi.org/10.1121/10.0025508","url":null,"abstract":"A noise-insensitive cost function was developed for estimating the speed of harmonic acoustic sources in uniform linear motion. This function weighs and integrates the energy distribution of received tones in the time-frequency plane to enhance the robustness of parameter estimation under low signal-to-noise ratio conditions, where weight values are intentionally combined with the law of observed instantaneous frequency. As the cost function is differentiable, the procedure of parameter estimations also has high computing efficiency. Processing data of SWellEx-96 experiments with real ocean noise confirmed the anti-noise capabilities of this cost function to conventional processing methods.","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140791512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A data-driven assessment of harmony in Quebec French [e] and [ε].","authors":"Josiane Riverin-Coutlée, Michele Gubian","doi":"10.1121/10.0025831","DOIUrl":"https://doi.org/10.1121/10.0025831","url":null,"abstract":"<p><p>This study is concerned with the aperture of the mid vowel /E/ in nonfinal syllables in Quebec French. The hypothesis tested is that in underived disyllabic words, the aperture of /E/ would be determined via harmony with the following vowel. Based on predictions from a classifier trained on acoustic properties of word-final vowels, nonfinal vowels were labeled as mid-close or mid-open. Although distant coarticulatory effects were observed, the harmony hypothesis was not supported. The results revealed a bias toward a mid-open quality and a reduced acoustic distinction, which warrant further investigation.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140867298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}