{"title":"Mouth rhythm as a \"packaging mechanism\" of information in speech: A proof of concept.","authors":"Lei He","doi":"10.1121/10.0035944","DOIUrl":"https://doi.org/10.1121/10.0035944","url":null,"abstract":"<p><p>This paper postulated and tested the possibility that the mouth rhythm functions as a \"packaging mechanism\" of information in speech. Cross-spectral analysis between two time series of mouth aperture size [parameterized as sample-by-sample interlip distances, i.e., o(t)] and information variations in speech [parameterized as frame-by-frame spectral entropy values, i.e., h(t)] was employed to reveal their underlying spectro-temporal relationship. Using a corpus containing more than 1000 utterances produced by a typical British English speaker, it was observed that both signals share slow recurring rates corresponding to the stress and syllable, with a slight phase lag of h(t) behind o(t) in the vicinity of 5 Hz.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"1612-1617"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143542428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Patti Adank, Han Wang, Taylor Hepworth, Stephanie A Borrie
{"title":"Perceptual adaptation to dysarthric speech is modulated by concurrent phonological processing: A dual task study.","authors":"Patti Adank, Han Wang, Taylor Hepworth, Stephanie A Borrie","doi":"10.1121/10.0035883","DOIUrl":"10.1121/10.0035883","url":null,"abstract":"<p><p>Listeners can adapt to noise-vocoded speech under divided attention using a dual task design [Wang, Chen, Yan, McGettigan, Rosen, and Adank, Trends Hear. 27, 23312165231192297 (2023)]. Adaptation to noise-vocoded speech, an artificial degradation, was largely unaffected for domain-general (visuomotor) and domain-specific (semantic or phonological) dual tasks. The study by Wang et al. was replicated in an online between-subject experiment with 4 conditions (N = 192) using 40 dysarthric sentences, a natural, real-world variation of the speech signal listeners can adapt to, to provide a closer test of the role of attention in adaptation. Participants completed a speech-only task (control) or a dual task, aiming to recruit domain-specific (phonological or lexical) or domain-general (visual) attentional processes. The results showed initial suppression of adaptation in the phonological condition during the first ten trials in addition to poorer overall speech comprehension compared to the speech-only, lexical, and visuomotor conditions. Yet, as there was no difference in the rate of adaptation across the 40 trials for the 4 conditions, it was concluded that perceptual adaptation to dysarthric speech could occur under divided attention, and it seems likely that adaptation is an automatic cognitive process that can occur under load.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"1598-1611"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11905114/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143586048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mingrui Liu, James W Wiskin, Gregory J Czarnota, Michael L Oelze
{"title":"Angular spatial compounding of diffraction corrected images improves ultrasound attenuation measurements.","authors":"Mingrui Liu, James W Wiskin, Gregory J Czarnota, Michael L Oelze","doi":"10.1121/10.0036124","DOIUrl":"10.1121/10.0036124","url":null,"abstract":"<p><p>Breast cancer is a leading cause of death for women. Quantitative ultrasound (QUS) and ultrasound computed tomography (USCT) are quantitative imaging techniques that have been investigated for management of breast cancer. QUS and USCT can generate ultrasound attenuation images. In QUS, the spectral log difference (SLD) is a technique that can provide estimates of the attenuation coefficient slope. Full angular spatial compounding (FASC) can be used with SLD to generate attenuation maps with better spatial resolution and lower estimate variance. In USCT, high quality speed of sound (SOS) images can be generated using full wave inversion (FWI) method, but attenuation images created using FWI are often of inferior quality. With the QTI Breast Acoustic CTTM Scanner (QT Imaging, Inc., Novato, CA), raw in-phase and quadrature data were used to implement SLD combined with FASC. The capabilities of SLD were compared with FWI through simulations, phantom experiments, and in vivo breast experiments. Results show the SLD resulted in improved accuracy in estimating lesion sizes compared to FWI. Further, SLD images had lower variance and mean absolute error (MAE) compared to FWI of the same samples with respect to the attenuation values (reducing MAE by three times) in the tissue mimicking phantoms.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"1638-1649"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11890159/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143575703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rouben Rehman, Christian Dreier, Jonas Heck, Josep Llorca-Bofí, Michael Vorländer
{"title":"Comparison of virtual reality and web-based listening experiments on the perception in complex auralized environmentsa).","authors":"Rouben Rehman, Christian Dreier, Jonas Heck, Josep Llorca-Bofí, Michael Vorländer","doi":"10.1121/10.0036147","DOIUrl":"https://doi.org/10.1121/10.0036147","url":null,"abstract":"<p><p>Listening experiments are crucial for understanding human sound perception. In overall human perception, combined audiovisual effects play an important role. However, traditional virtual reality (VR) setups, consisting of a head-mounted display (HMD) and headphones, are limited by their need for expensive equipment and time-consuming laboratory sessions. Striving for alternatives, online experiments have demonstrated their potential in other areas of research. However, these experiments have been restricted to basic setups lacking interactivity. This study presents a web-based approach with audiovisual experiments being run on a server and streamed in real time. To this end, two reproduction setups are compared: an immersive laboratory setup (HMD-based visualization with controller navigation and headphones) and a consumer setup (screen-based visualization with keyboard navigation and headphones). The experiment comprises quality ratings and noise assessments of four auralized noise conditions with additional visualization. For noise perception experiments, the results are promising, showing minimal differences in questionnaire ratings between VR and streaming reproduction. Visual quality ratings suffered mildly in the consumer setup, but auralization quality was rated similarly positive in both cases. Even for lower feeling of presence in the consumer setup, the subjects' attention remained similarly high. Finally, accessibility and quality ratings indicate promising results, too.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"2001-2017"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143692505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dorian S Houser, Kyle Donohoe, Jason Mulsow, James J Finneran
{"title":"Quantifying differences in dolphin hearing thresholds obtained with behavioral and auditory evoked potential methods.","authors":"Dorian S Houser, Kyle Donohoe, Jason Mulsow, James J Finneran","doi":"10.1121/10.0036153","DOIUrl":"https://doi.org/10.1121/10.0036153","url":null,"abstract":"<p><p>Different methods of producing the auditory steady state response (ASSR) are used to test dolphin hearing, but each method affects the resulting ASSR threshold. Since behavioral thresholds are often desired, this study, using common ASSR methods, compared differences between ASSR and behavioral hearing thresholds in five dolphins. Sinusoidal amplitude modulated (SAM) tones or tone pip trains were presented to the dolphins through a contact transducer while they were in air or partially submerged under water. Underwater behavioral hearing thresholds were obtained with pure tone stimuli on the same days as ASSR testing. Independent of the test medium, SAM tone stimuli yielded thresholds that consistently overestimated (i.e., were higher than) behavioral thresholds. Tone pip trains consistently underestimated thresholds when presented in air, and while they underestimated thresholds at lower test frequencies, they overestimated thresholds at higher test frequencies when presented under water. The mean differences between ASSR and behavioral thresholds were almost always lower when using tone pip train stimuli, but were exaggerated up to -47 dB when testing frequencies just above the upper-frequency limit of hearing. Knowing the relationship between ASSR and behavioral thresholds enables better approximations of behavioral thresholds in dolphins for which only ASSR thresholds exist.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"1955-1968"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143674221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Talker-variability in visual feedback articulatory training: Transfer to words, long-term maintenance, and generalization.","authors":"Natalia Kartushina, Clara D Martin","doi":"10.1121/10.0036217","DOIUrl":"https://doi.org/10.1121/10.0036217","url":null,"abstract":"<p><p>An important challenge in second language acquisition is learning how to produce novel sounds. Visual-acoustic feedback training has been shown to improve second language production, but it is still unclear whether training on isolated sounds transfers to words and is maintained over time. Forty-six Spanish speakers were trained to produce unfamiliar French vowels /e/-/ɛ/ through exposure to either single- or multiple-talker stimuli. On each trial, they heard a vowel, produced it, and received feedback showing the position of their vowel and the target in the acoustic space. Both groups showed improvements in vowel production, which transferred to words and were maintained 12 days later. However, better generalization-more accurate production when repeating vowels after an unfamiliar speaker-was maintained only in the multiple-talker group. This study shows that visual-acoustic feedback training is an effective method for improving second language production as the gains of such training demonstrate three key features for pedagogical practice: transfer, generalization, and maintenance.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"2117-2126"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143709918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jonathan Regev, Johannes Zaar, Helia Relaño-Iborra, Torsten Dau
{"title":"Investigating the effects of age and hearing loss on speech intelligibility and amplitude modulation frequency selectivity.","authors":"Jonathan Regev, Johannes Zaar, Helia Relaño-Iborra, Torsten Dau","doi":"10.1121/10.0036220","DOIUrl":"https://doi.org/10.1121/10.0036220","url":null,"abstract":"<p><p>The perception of amplitude modulation (AM), characterized by a frequency-selective process in the modulation domain, is considered critical for speech intelligibility. Previous literature has provided evidence of an age-related decline in AM frequency selectivity. Additionally, a notable sharpening of AM tuning with hearing loss has been observed, which was proposed to be related to a perceptual advantage resulting from peripheral compression loss. This study explores whether such changes in AM tuning with age and hearing loss contribute to the speech intelligibility challenges older listeners often face in noisy environments. Young (n = 10, aged 22-28 years) and older (n = 9, aged 57-77 years) listeners with normal hearing as well as older listeners with hearing impairment (n = 9, aged 64-77 years) participated in the study. All had previously taken part in studies on AM tuning. Speech-reception thresholds (SRTs) were collected under conditions including stationary, fluctuating, and competing-speech maskers. The results revealed an age-related increase in SRTs, with an additional negative impact of hearing loss. Beyond age and audibility, a measure of AM tuning significantly contributed to explaining the variance in SRTs across listeners for specific maskers. These findings motivate further exploration of the relationship between AM frequency selectivity and speech intelligibility in noise.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"2077-2090"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143700700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Michael G T Wikinson, XingYao York Wang, Noah J Cowan, Cynthia F Moss
{"title":"Echolocating bats adjust sonar call features and head/ear position as they track moving targets in the presence of clutter.","authors":"Michael G T Wikinson, XingYao York Wang, Noah J Cowan, Cynthia F Moss","doi":"10.1121/10.0036252","DOIUrl":"https://doi.org/10.1121/10.0036252","url":null,"abstract":"<p><p>Echolocating bats often encounter clutter as they pursue insect prey. To probe the adaptive behaviors bats employ to mitigate the effects of clutter, this study quantified echolocation call features and head movements of big brown bats (Eptesicus fuscus) as they tracked a moving prey target in the dark. Bats were trained to rest on a perch and track an approaching target for a food reward. Clutter was positioned at different distances and angular offsets from the bat and the path of a moving target. This study hypothesized that bats dynamically adjust call features and head direction to facilitate target localization in the presence of clutter. The results show that bats shortened call duration and interval and increased head movements when the target was close to clutter. The study also revealed that bats increase the production of sonar strobe groups in cluttered environments, which may sharpen sonar spatial resolution. Spectral analysis showed that maximum call power shifted to lower frequencies when clutter was close to the target. These data demonstrate the big brown bat's range of adaptive behaviors that support target tracking in cluttered environments.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"2236-2247"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143719925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ruigang Han, Ning Jia, Yunfei Li, Dong Xiao, Shengming Guo, Li Ma
{"title":"Iterative-detection-based time-domain adaptive decision feedback equalization for continuous phase modulation of underwater acoustic communication.","authors":"Ruigang Han, Ning Jia, Yunfei Li, Dong Xiao, Shengming Guo, Li Ma","doi":"10.1121/10.0036145","DOIUrl":"https://doi.org/10.1121/10.0036145","url":null,"abstract":"<p><p>Continuous phase modulation (CPM), which is widely used in aviation telemetry and satellite communications, may help improve the performance of underwater acoustic (UWA) communication systems owing to its high spectral and power efficiency. However, applying conventional frequency-domain equalization (FDE) algorithms to CPM signals over time-varying UWA channels considerably degrades performance. Moreover, time-domain equalization algorithms often rely on excessive approximations for symbol detection, compromising overall reception. This study presents an iterative-detection-based time-domain adaptive decision feedback equalization (ID-TDADFE) algorithm that tracks channel variations through symbol-by-symbol detection. The symbol detection in ID-TDADFE fully considers the inherent coding gain of CPM signals can be cascaded with an adaptive equalizer, and enhances symbol detection performance by utilizing joint probability estimation. Numerical simulations with minimum-shift keying (MSK) and Gaussian MSK signals demonstrated that ID-TDADFE significantly improved communication performance over a time-varying UWA channel within one or two iterations. In a sea trial for experimental verification, ID-TDADFE reduced bit errors by 45.08% and 51.8% in the first and second iterations, respectively, compared to FDE.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"1912-1925"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143657579","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kevin M Lee, Kelly M Dorgan, Gabriel R Venegas, Jason D Chaytor, Megan S Ballard, Andrew R McNeese, Preston S Wilson
{"title":"Investigation of surficial seabed heterogeneity and geoacoustic variability in the New England Mud Patcha).","authors":"Kevin M Lee, Kelly M Dorgan, Gabriel R Venegas, Jason D Chaytor, Megan S Ballard, Andrew R McNeese, Preston S Wilson","doi":"10.1121/10.0036122","DOIUrl":"https://doi.org/10.1121/10.0036122","url":null,"abstract":"<p><p>Benthic biological processes influence seabed heterogeneity and contribute to variability in geoacoustic properties. To investigate these relationships, measurements were conducted to quantify spatial variability in the upper few decimeters of sediment near the water-seabed interface within a fine-grained sediment deposit on the New England continental shelf. At each measurement location, an acoustic multicorer was deployed to sample the seabed. Acoustic probes were inserted into the sediment to collect direct in situ measurements of sediment compressional wave speed and attenuation (30-100 kHz) under near-ambient conditions, after which cores were collected from the inter-probe propagation paths. Sediment physical properties, organic carbon, infaunal community composition, and ex situ compressional wave speed and attenuation spanning two frequency decades (104-106 Hz) were subsequently measured in the laboratory. The frequency dependence of sound speed ratio and attenuation was analyzed in the context of sediment acoustics models for mud based on the viscous grain shearing and extended Biot models. Sites with greater abundance of larger-bodied infauna (>1 mm) displayed higher variability in sound speed and attenuation. Correlation was found between sediment compressional wave modulus and total organic carbon, suggesting that organic matter in the sediment matrix also affects bulk acoustic properties.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"157 3","pages":"1686-1702"},"PeriodicalIF":2.1,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143615716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}