Christophe Lesimple, Volker Kuehnel, Kai Siedenburg
{"title":"Hearing aid evaluation for music: Accounting for acoustical variability of music stimuli.","authors":"Christophe Lesimple, Volker Kuehnel, Kai Siedenburg","doi":"10.1121/10.0028397","DOIUrl":"10.1121/10.0028397","url":null,"abstract":"<p><p>Music is an important signal class for hearing aids, and musical genre is often used as a descriptor for stimulus selection. However, little research has systematically investigated the acoustical properties of musical genres with respect to hearing aid amplification. Here, extracts from a combination of two comprehensive music databases were acoustically analyzed. Considerable overlap in acoustic descriptor space between genres emerged. By simulating hearing aid processing, it was shown that effects of amplification regarding dynamic range compression and spectral weighting differed across musical genres, underlining the critical role of systematic stimulus selection for research on music and hearing aids.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Comparing human and machine speech recognition in noise with QuickSIN.","authors":"Malcolm Slaney, Matthew B Fitzgerald","doi":"10.1121/10.0028612","DOIUrl":"10.1121/10.0028612","url":null,"abstract":"<p><p>A test is proposed to characterize the performance of speech recognition systems. The QuickSIN test is used by audiologists to measure the ability of humans to recognize continuous speech in noise. This test yields the signal-to-noise ratio at which individuals can correctly recognize 50% of the keywords in low-context sentences. It is argued that a metric for automatic speech recognizers will ground the performance of automatic speech-in-noise recognizers to human abilities. Here, it is demonstrated that the performance of modern recognizers, built using millions of hours of unsupervised training data, is anywhere from normal to mildly impaired in noise compared to human participants.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142156849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A method for calculating the grand average of a set of auditory brain-stem responses.","authors":"Sinnet G B Kristensen, Claus Elberling","doi":"10.1121/10.0028320","DOIUrl":"10.1121/10.0028320","url":null,"abstract":"<p><p>To calculate a grand average waveform for a set of auditory brain-stem responses (ABRs), no generally accepted method exists. Here, we evaluate a new method using temporal adjustment of the underlying ABRs. Compared to a method without temporal adjustment, the new method results in higher amplitudes of the individual waves in the grand average. The grand average produced by the method better represents the group mean wave-amplitudes because it reduces smearing of the individual waves caused by inter-subject latency variability.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Joshua G W Bernstein, Julianna Voelker, Sandeep A Phatak
{"title":"Headphones over the cochlear-implant sound processor to replace direct audio input.","authors":"Joshua G W Bernstein, Julianna Voelker, Sandeep A Phatak","doi":"10.1121/10.0028737","DOIUrl":"10.1121/10.0028737","url":null,"abstract":"<p><p>Psychoacoustic stimulus presentation to the cochlear implant via direct audio input (DAI) is no longer possible for many newer sound processors (SPs). This study assessed the feasibility of placing circumaural headphones over the SP. Calibration spectra for loudspeaker, DAI, and headphone modalities were estimated by measuring cochlear-implant electrical output levels for tones presented to SPs on an acoustic manikin. Differences in calibration spectra between modalities arose mainly from microphone-response characteristics (high-frequency differences between DAI and the other modalities) or a proximity effect (low-frequency differences between headphones and loudspeaker). Calibration tables are provided to adjust for differences between the three modalities.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hyeonjeong Park, Fangxu Xing, Maureen Stone, Hahn Kang, Xiaofeng Liu, Jiachen Zhuo, Sidney Fels, Timothy G Reese, Van J Wedeen, Georges El Fakhri, Jerry L Prince, Jonghye Woo
{"title":"Investigating muscle coordination patterns with Granger causality analysis in protrusive motion from tagged and diffusion MRI.","authors":"Hyeonjeong Park, Fangxu Xing, Maureen Stone, Hahn Kang, Xiaofeng Liu, Jiachen Zhuo, Sidney Fels, Timothy G Reese, Van J Wedeen, Georges El Fakhri, Jerry L Prince, Jonghye Woo","doi":"10.1121/10.0028500","DOIUrl":"10.1121/10.0028500","url":null,"abstract":"<p><p>The human tongue exhibits an orchestrated arrangement of internal muscles, working in sequential order to execute tongue movements. Understanding the muscle coordination patterns involved in tongue protrusive motion is crucial for advancing knowledge of tongue structure and function. To achieve this, this work focuses on five muscles known to contribute to protrusive motion. Tagged and diffusion MRI data are collected for analysis of muscle fiber geometry and motion patterns. Lagrangian strain measurements are derived, and Granger causal analysis is carried out to assess predictive information among the muscles. Experimental results suggest sequential muscle coordination of protrusive motion among distinct muscle groups.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11384280/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142141922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Monaural and binaural masking release with speech-like stimuli.","authors":"Hyojin Kim, Viktorija Ratkute, Bastian Epp","doi":"10.1121/10.0028736","DOIUrl":"10.1121/10.0028736","url":null,"abstract":"<p><p>The relevance of comodulation and interaural phase difference for speech perception is still unclear. We used speech-like stimuli to link spectro-temporal properties of formants with masking release. The stimuli comprised a tone and three masker bands centered at formant frequencies F1, F2, and F3 derived from a consonant-vowel. The target was a diotic or dichotic frequency-modulated tone following F2 trajectories. Results showed a small comodulation masking release, while the binaural masking level difference was comparable to previous findings. The data suggest that factors other than comodulation may play a dominant role in grouping frequency components in speech.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Minima in cubic distortion-product otoacoustic emission input/output functions due to distributed primary sources.","authors":"Václav Vencovský, Aleš Vetešník","doi":"10.1121/10.0028386","DOIUrl":"https://doi.org/10.1121/10.0028386","url":null,"abstract":"<p><p>Input/output (I/O) functions of distortion-product otoacoustic emissions (DPOAEs) may contain sudden amplitude minima (notches) although they are measured in animals with a negligible reflection source. We measured DPOAEs in humans and analyzed the data by time-frequency filtering to decompose the nonlinear-distortion and coherent-reflection components of DPOAE. The presented I/O functions of the nonlinear-distortion component contain notches. We suggest that because these notches are present only in the nonlinear-distortion component, they result from destructive interference between distortion-product wavelets coming from the primary generation region. Simulations conducted with a nonlinear cochlear model showed qualitative similarities with the presented experimental results.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Individual variability in the use of tonal and non-tonal cues in intonationa).","authors":"Na Hu, Amalia Arvaniti","doi":"10.1121/10.0028613","DOIUrl":"10.1121/10.0028613","url":null,"abstract":"<p><p>Greek uses H*, L + H*, and H* + L, all followed by L-L% edge tones, as nuclear pitch accents in statements. A previous analysis demonstrated that these accents are distinguished by F0 scaling and contour shape. This study expands the earlier investigation by exploring additional cues, namely, voice quality, amplitude, and duration, in distinguishing the pitch accents, and investigating individual variability in the selection of both F0 and non-F0 cues. Bayesian multivariate analysis and hierarchical clustering demonstrate that the accents are distinguished not only by F0 but also by additional cues at the group level, with individual variability in cue selection.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Effect of hearing aids on the externalization of everyday sounds.","authors":"Elin Roverud, Virginia Best","doi":"10.1121/10.0028381","DOIUrl":"10.1121/10.0028381","url":null,"abstract":"<p><p>This study examined the influence of stimulus properties on sound externalization when listening with hearing aids. Normally hearing listeners were presented with broadband \"tokens\" (environmental sounds and speech) from loudspeakers, and rated externalization using a continuous scale. In separate blocks, they listened unaided or while wearing behind-the-ear hearing aids with closed domes and low gain (linear or compressive). There was a significant influence of token on ratings, even for unaided listening, and the effect of hearing aids depended on token. An acoustic analysis indicated that hearing aids were more likely to disrupt externalization for peakier sounds with a low-frequency emphasis.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11382285/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142121269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Karolina Prawda, Sebastian J Schlecht, Vesa Välimäki
{"title":"Non-stationary noise removal from repeated sweep measurements.","authors":"Karolina Prawda, Sebastian J Schlecht, Vesa Välimäki","doi":"10.1121/10.0028203","DOIUrl":"https://doi.org/10.1121/10.0028203","url":null,"abstract":"<p><p>Acoustic measurements using sine sweeps are prone to background noise and non-stationary disturbances. Repeated measurements can be averaged to improve the resulting signal-to-noise ratio. However, averaging leads to poor rejection of non-stationary high-energy disturbances and, in the case of a time-variant environment, causes attenuation at high frequencies. This paper proposes a robust method to combine repeated sweep measurements using across-measurement median filtering in the time-frequency domain. The method, called Mosaic, successfully rejects non-stationary noise, suppresses background noise, and is more robust toward time variation than averaging. The proposed method allows high-quality measurement of impulse responses in a noisy environment.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141986719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}