Xueli Sheng, Dewen Li, Ran Cao, Xuan Zhou, Jiarui Yin
{"title":"Adaptive interference suppression based on an invariant subspace of matrices matching for a horizontal array in underwater acoustics.","authors":"Xueli Sheng, Dewen Li, Ran Cao, Xuan Zhou, Jiarui Yin","doi":"10.1121/10.0026373","DOIUrl":"https://doi.org/10.1121/10.0026373","url":null,"abstract":"<p><p>Passive detection of target-of-interest (TOI) within strong interferences poses a challenge. This paper introduces an adaptive interference suppression based on an invariant subspace of matrix matching. Assume that the TOI-bearing intervals are known. We define a correlation ratio for each eigenvector to obtain the highest one. Then, we use invariant subspace of matrix matching to measure the distance between two invariant projection matrices of this eigenvector. This identifies and removes the eigenvectors associated with TOI. Finally, the remaining eigenvectors are subtracted from the sample covariance matrix to suppress interference and noise. The viability of the proposed method is demonstrated experimentally.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141312447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Joint equalization and decoding with per-survivor processing based on super-trellis in time-varying underwater acoustic channels.","authors":"Xu Kou, Yanbo Wu, Min Zhu","doi":"10.1121/10.0026372","DOIUrl":"https://doi.org/10.1121/10.0026372","url":null,"abstract":"<p><p>This Letter proposes a low-complexity joint equalization and decoding reception scheme based on super-trellis per-survivor processing, making it possible to apply maximum likelihood sequence estimation in high-order underwater acoustic communications under fast time-varying channels. The technique combines trellis-coded modulation states and intersymbol interference states and uses per-survivor processing to track channel parameters. Furthermore, a general trellis configuration for arbitrary order quadrature amplitude modulation signal is provided when truncate the channel is used to describe the intersymbol interference state to 1. Sea trials results show that the performance of proposed method can be more than 1.4 dB superiority than conventional schemes.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141319167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Neural correlates of tonal loudness, intensity discrimination, and duration discrimination.","authors":"Shigeyuki Kuwada, Constantine Trahiotis","doi":"10.1121/10.0025874","DOIUrl":"10.1121/10.0025874","url":null,"abstract":"<p><p>A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Static and moving minimum audible angle: Independent contributions of reverberation and position.","authors":"Anna Dietze, Samuel W Clapp, Bernhard U Seeber","doi":"10.1121/10.0025992","DOIUrl":"https://doi.org/10.1121/10.0025992","url":null,"abstract":"<p><p>Two measures of auditory spatial resolution, the minimum audible angle and the minimum audible movement angle, have been obtained in a simulated acoustic environment using Ambisonics sound field reproduction. Trajectories were designed to provide no reliable cues for the spatial discrimination task. Larger threshold angles were found in reverberant compared to anechoic conditions, for stimuli on the side compared to the front, and for moving compared to static stimuli. The effect of reverberation appeared to be independent of the position of the sound source (same relative threshold increase) and was independently present for static and moving sound sources.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140923846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima
{"title":"Impact of speech rate on perception of vowel and consonant duration by bilinguals and monolinguals.","authors":"Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima","doi":"10.1121/10.0025862","DOIUrl":"10.1121/10.0025862","url":null,"abstract":"<p><p>The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Predicting underwater acoustic transmission loss in the SOFAR channel from ray trajectories via deep learning.","authors":"Haitao Wang, Shiwei Peng, Qunyi He, Xiangyang Zeng","doi":"10.1121/10.0025976","DOIUrl":"https://doi.org/10.1121/10.0025976","url":null,"abstract":"<p><p>Predicting acoustic transmission loss in the SOFAR channel faces challenges, such as excessively complex algorithms and computationally intensive calculations in classical methods. To address these challenges, a deep learning-based underwater acoustic transmission loss prediction method is proposed. By properly training a U-net-type convolutional neural network, the method can provide an accurate mapping between ray trajectories and the transmission loss over the problem domain. Verifications are performed in a SOFAR channel with Munk's sound speed profile. The results suggest that the method has potential to be used as a fast predicting model without sacrificing accuracy.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya
{"title":"Reduced digit spans and ear dominance using dichotic digits in bimodal cochlear-implant users.","authors":"Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya","doi":"10.1121/10.0025977","DOIUrl":"10.1121/10.0025977","url":null,"abstract":"<p><p>Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11550484/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140899058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An open auscultation dataset for machine learning-based respiratory diagnosis studies.","authors":"Guanyu Zhou, Chengjian Liu, Xiaoguang Li, Sicong Liang, Ruichen Wang, Xun Huang","doi":"10.1121/10.0025851","DOIUrl":"10.1121/10.0025851","url":null,"abstract":"<p><p>Machine learning enabled auscultating diagnosis can provide promising solutions especially for prescreening purposes. The bottleneck for its potential success is that high-quality datasets for training are still scarce. An open auscultation dataset that consists of samples and annotations from patients and healthy individuals is established in this work for the respiratory diagnosis studies with machine learning, which is of both scientific importance and practical potential. A machine learning approach is examined to showcase the use of this new dataset for lung sound classifications with different diseases. The open dataset is available to the public online.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Clear speech effects in production of sentence-medial Mandarin lexical tonesa).","authors":"Jack Rittenberry, Irina A Shport","doi":"10.1121/10.0025991","DOIUrl":"10.1121/10.0025991","url":null,"abstract":"<p><p>Adding to limited research on clear speech in tone languages, productions of Mandarin lexical tones were examined in pentasyllabic sentences. Fourteen participants read sentences imagining a hard-of-hearing addressee or a friend in a casual social setting. Tones produced in clear speech had longer duration, higher intensity, and larger F0 values. This style effect was rarely modulated by tone, preceding tonal context, or syllable position, consistent with an overall signal enhancement strategy. Possible evidence for tone enhancement was observed only in one set of analysis for F0 minimum and F0 range, contrasting tones with low targets and tones with high targets.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141158923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Amara C Ezenwa, Matthew J Goupell, Sandra Gordon-Salant
{"title":"Cochlear-implant listeners benefit from training with time-compressed speech, even at advanced ages.","authors":"Amara C Ezenwa, Matthew J Goupell, Sandra Gordon-Salant","doi":"10.1121/10.0025431","DOIUrl":"10.1121/10.0025431","url":null,"abstract":"<p><p>This study evaluated whether adaptive training with time-compressed speech produces an age-dependent improvement in speech recognition in 14 adult cochlear-implant users. The protocol consisted of a pretest, 5 h of training, and a posttest using time-compressed speech and an adaptive procedure. There were significant improvements in time-compressed speech recognition at the posttest session following training (>5% in the average time-compressed speech recognition threshold) but no effects of age. These results are promising for the use of adaptive training in aural rehabilitation strategies for cochlear-implant users across the adult lifespan and possibly using speech signals, such as time-compressed speech, to train temporal processing.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11075136/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}