Andrey G Tyshchenko, Mikhail A Sorokin, Sergey B Kozitskiy, Pavel S Petrov
{"title":"The solution of sound propagation modeling problems for environment impact assessment by the mode parabolic equations methoda).","authors":"Andrey G Tyshchenko, Mikhail A Sorokin, Sergey B Kozitskiy, Pavel S Petrov","doi":"10.1121/10.0034424","DOIUrl":"10.1121/10.0034424","url":null,"abstract":"<p><p>The method of sound propagation modeling based on the mode parabolic equations (MPEs) theory is applied to the verification scenarios for environmental impact assessment. The results for selected scenarios from the 2022 Cambridge Joint Industry Programme Acoustic Modelling Workshop and the configuration of the computational programs AMPLE and MPE for these scenarios is discussed. Furthermore, it is revealed how the results for these scenarios change in the case of the bottom slope across and along the propagation path. It is observed that for the cross-slope propagation scenario, the distribution of acoustic energy over decidecade frequency bands does not depend on the slope angle and is practically the same as that for range-independent environment. At the same time, the dependence of energy distribution is noticeable for up- and downslope propagation scenarios, where greater slope angles result in higher propagation loss. It is also shown that MPEs are capable of adequately handling typical sound propagation problems related to the environmental impact assessment for frequencies up to 1000 Hz. A possibility of using frequency-dependent mesh size and number of modes must be implemented in codes based on this approach.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3306-3319"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142638849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hongyan Zhang, Linfeng Wang, Xin Chen, Jian Li, Yiwei Liu, Haichao Liu, Yang Liu
{"title":"On propagation characteristics of ultrasonic guided waves in layered fluid-saturated porous media using spectral method.","authors":"Hongyan Zhang, Linfeng Wang, Xin Chen, Jian Li, Yiwei Liu, Haichao Liu, Yang Liu","doi":"10.1121/10.0034232","DOIUrl":"10.1121/10.0034232","url":null,"abstract":"<p><p>Fluid-saturated porous media plays an increasingly important role in emerging fields such as lithium batteries and artificial bones. Accurately solving the governing equations of guided wave is the key to the successful application of ultrasonic guided wave nondestructive testing technology in fluid-saturated porous media. This paper derives the Lamb wave equation in layered fluid-saturated porous materials based on Biot theory and proposes the spectral method suitable for solving complex wave equations. The spectral method reconstructs the fundamental wave equations in the form of a matrix eigenvalue problem using spectral differentiation matrices. It introduces boundary conditions by replacing corresponding rows in the wave equation matrix with stress or displacement in matrix form. For complex differential equations, such as the governing equations of guided waves in porous media, the spectral method has the significant advantages of faster computation speed, less root loss, and easier encoding process. The spectral method is used to calculate the acoustic field characteristics under different boundary conditions and environments of the layer fluid-saturated porous media. Results show that the surface treatment details and environment of fluid-saturated porous materials play an important role in the propagation of guided waves.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3021-3032"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142568997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Sound source locations and their roles in Japanese voiceless \"glottal\" fricative production.","authors":"Tsukasa Yoshinaga, Kikuo Maekawa, Akiyoshi Iida","doi":"10.1121/10.0034229","DOIUrl":"10.1121/10.0034229","url":null,"abstract":"<p><p>Although [h] is described as a glottal fricative, it has never been demonstrated whether [h] has its source exclusively at the glottis. In this study, sound source locations and their influence on sound amplitudes were investigated by conducting mechanical experiments and airflow simulations. Vocal tract data of [h] were obtained in three phonemic contexts from two native Japanese subjects using three-dimensional static magnetic resonance imaging (MRI). Acrylic vocal tract replicas were constructed, and the sound was reproduced by supplying airflow to the vocal tracts with adducted or abducted vocal folds. The sound source locations were estimated by solving the Navier-Stokes equations. The results showed that the amplitudes of sounds produced by the vocal tracts with an open glottis were in a similar range (±3 dB) to those with a glottal gap of 3 mm in some contexts. The sound sources in these cases were observed in the pharyngeal cavity or near the soft palate. Similar degrees of oral constrictions were observed in the real-time MRI, indicating that the sound traditionally described as [h] is produced, at least in some contexts, with sound sources of turbulent flow generated by a supralaryngeal constriction of the following vowel.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"2935-2948"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142558112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Florian Kattner, Julia Föcker, Cleopatra Christina Moshona, John E Marsh
{"title":"When softer sounds are more distracting: Task-irrelevant whispered speech causes disruption of serial recall.","authors":"Florian Kattner, Julia Föcker, Cleopatra Christina Moshona, John E Marsh","doi":"10.1121/10.0034454","DOIUrl":"https://doi.org/10.1121/10.0034454","url":null,"abstract":"<p><p>Two competing accounts propose that the disruption of short-term memory by irrelevant speech arises either due to interference-by-process (e.g., changing-state effect) or attentional capture, but it is unclear how whispering affects the irrelevant speech effect. According to the interference-by-process account, whispered speech should be less disruptive due to its reduced periodic spectro-temporal fine structure and lower amplitude modulations. In contrast, the attentional account predicts more disruption by whispered speech, possibly via enhanced listening effort in the case of a comprehended language. In two experiments, voiced and whispered speech (spoken sentences or monosyllabic words) were presented while participants memorized the order of visually presented letters. In both experiments, a changing-state effect was observed regardless of the phonation (sentences produced more disruption than \"steady-state\" words). Moreover, whispered speech (lower fluctuation strength) was more disruptive than voiced speech when participants understood the language (Experiment 1), but not when the language was incomprehensible (Experiment 2). The results suggest two functionally distinct mechanisms of auditory distraction: While changing-state speech causes automatic interference with seriation processes regardless of its meaning or intelligibility, whispering appears to contain cues that divert attention from the focal task primarily when presented in a comprehended language, possibly via enhanced listening effort.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3632-3648"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142716477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
W Erick Rogers, Laurie T Fialkowski, Daniel J Brooker, Gleb Panteleev, Joseph M Fialkowski
{"title":"Utility of ocean wave parameters in ambient noise predictiona).","authors":"W Erick Rogers, Laurie T Fialkowski, Daniel J Brooker, Gleb Panteleev, Joseph M Fialkowski","doi":"10.1121/10.0034437","DOIUrl":"https://doi.org/10.1121/10.0034437","url":null,"abstract":"<p><p>This study is concerned with prediction of the \"wind noise\" component of ambient noise (AN) in the ocean. It builds on the seminal paper by Felizardo and Melville [(1995). J. Phys. Oceanogr. 25, 513-532], in which the authors quantified the correlation between AN and individual wind/wave parameters. Acoustic data are obtained from hydrophones at six diverse locations, and wind/wave parameters are obtained from moored buoys and numerical models. We describe a procedure developed for this study that identifies correlation of AN with wave parameters, independent of their mutual correlation with wind speed. We then describe paired calibration/prediction experiments, whereby multiple wind/wave parameters are used simultaneously to estimate AN. We find that the improvement from inclusion of wave parameters is robust but marginal; typically, root mean square error (RMSE) is reduced by less than 0.3 dB and/or less than 12% of the original RMSE. We interpret the latter outcome as suggesting that wave breaking responds to changes in local winds quickly, relative to, for example, total wave energy, which develops more slowly. This outcome is consistent with prior observations of wave breaking, e.g., Babanin [(2011). Breaking and Dissipation of Ocean Surface Waves (Cambridge University Press, Cambridge, UK), Chap. 3]. We discuss this in context of the time/space response of various wave parameters to wind forcing.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3649-3664"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142729937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hsin-Tien Chiang, Szu-Wei Fu, Hsin-Min Wang, Yu Tsao, John H L Hansen
{"title":"Multi-objective non-intrusive hearing-aid speech assessment model.","authors":"Hsin-Tien Chiang, Szu-Wei Fu, Hsin-Min Wang, Yu Tsao, John H L Hansen","doi":"10.1121/10.0034362","DOIUrl":"https://doi.org/10.1121/10.0034362","url":null,"abstract":"<p><p>Because a reference signal is often unavailable in real-world scenarios, reference-free speech quality and intelligibility assessment models are important for many speech processing applications. Despite a great number of deep-learning models that have been applied to build non-intrusive speech assessment approaches and achieve promising performance, studies focusing on the hearing impaired (HI) subjects are limited. This paper presents HASA-Net+, a multi-objective non-intrusive hearing-aid speech assessment model, building upon our previous work, HASA-Net. HASA-Net+ improves HASA-Net in several ways: (1) inclusivity for both normal-hearing and HI listeners, (2) integration with pre-trained speech foundation models and fine-tuning techniques, (3) expansion of predictive capabilities to cover speech quality and intelligibility in diverse conditions, including noisy, denoised, reverberant, dereverberated, and vocoded speech, thereby evaluating its robustness, and (4) validation of the generalization capability using an out-of-domain dataset.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3574-3587"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142716476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Manish Manohare, Francesco Aletta, Tin Oberman, Rajasekar Elangovan, Manoranjan Parida, Jian Kang
{"title":"Cross-country variation in psychophysiological responses to traffic noise exposure: Laboratory experiments in India and the UKa).","authors":"Manish Manohare, Francesco Aletta, Tin Oberman, Rajasekar Elangovan, Manoranjan Parida, Jian Kang","doi":"10.1121/10.0034242","DOIUrl":"https://doi.org/10.1121/10.0034242","url":null,"abstract":"<p><p>Traffic noise exposure has detrimental effects on human health, including both auditory and nonauditory impacts. As one such nonauditory factor, individuals and communities in different countries may exhibit different patterns of noise sensitivity and corresponding tolerance levels, leading to a change in overall noise perception. This paper investigated the cross-country differences in psychophysiological responses to traffic noise exposure between Indian and British individuals. A psychophysiological signal-based [heart rate variability (HRV) and skin conductance response (SCR)] listening experiment was conducted in India and the United Kingdom to analyze changes in noise perception and psychophysiological responses resulting from exposure to the same noise stimuli. HRV analysis indicated greater cardiovascular impact in the British group due to a significant increase in heart rate (W = 653, p < 0.01). Also, a significant increase in the SCR (W = 535, p < 0.001) was noted, indicating a greater level of physiological stress among British participants due to traffic noise stimuli. These findings highlight the difference in noise perception due to cross-country variation using psychophysiological responses. Understanding these cross-country differences can inform targeted interventions and policies to mitigate the adverse effects of traffic noise on human well-being.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3067-3079"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142583196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shuyang Jia, Lianglong Da, Sichen Zou, Baoheng Liu, Xiaochuan Zhang
{"title":"A D* orthogonal matching pursuit algorithm for time-varying channel estimation.","authors":"Shuyang Jia, Lianglong Da, Sichen Zou, Baoheng Liu, Xiaochuan Zhang","doi":"10.1121/10.0034367","DOIUrl":"https://doi.org/10.1121/10.0034367","url":null,"abstract":"<p><p>Orthogonal matching pursuit (OMP) combined with the A* search algorithm (A*OMP) exhibits robust reconstruction capabilities for synthesizing sparse data and signals, achieving relatively low reconstruction errors and a higher exact recovery ability than conventional OMP. However, A*OMP is only suitable for static channel estimation and cannot be applied to dynamic scenarios. This is because the channel delays for several consecutive orthogonal frequency-division multiplexing blocks per frame are similar and the path gains exhibit temporal correlation. This paper introduces a dynamic OMP approach (D*OMP) that employs a heuristic function of A*OMP and a unique reverse process, enabling sparse solutions to be identified in unknown and changing environments. The proposed method is highly practical for joint channel estimation across multiple blocks. Simulation and sea trial results indicate that D*OMP not only possesses superior channel recovery accuracy, but also has a more efficient channel reconstruction process, outperforming both A*OMP and conventional OMP.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3158-3168"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142622973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mingyue Huo, Yinglun Sun, Daniel Fogerty, Yan Tang
{"title":"Release from same-talker speech-in-speech masking: Effects of masker intelligibility and other contributing factorsa).","authors":"Mingyue Huo, Yinglun Sun, Daniel Fogerty, Yan Tang","doi":"10.1121/10.0034235","DOIUrl":"https://doi.org/10.1121/10.0034235","url":null,"abstract":"<p><p>Human speech perception declines in the presence of masking speech, particularly when the masker is intelligible and acoustically similar to the target. A prior investigation demonstrated a substantial reduction in masking when the intelligibility of competing speech was reduced by corrupting voiced segments with noise [Huo, Sun, Fogerty, and Tang (2023), \"Quantifying informational masking due to masker intelligibility in same-talker speech-in-speech perception,\" in Interspeech 2023, pp. 1783-1787]. As this processing also reduced the prominence of voiced segments, it was unclear whether the unmasking was due to reduced linguistic content, acoustic similarity, or both. The current study compared the masking of original competing speech (high intelligibility) to competing speech with time reversal of voiced segments (VS-reversed, low intelligibility) at various target-to-masker ratios. Modeling results demonstrated similar energetic masking between the two maskers. However, intelligibility of the target speech was considerably better with the VS-reversed masker compared to the original masker, likely due to the reduced linguistic content. Further corrupting the masker's voiced segments resulted in additional release from masking. Acoustic analyses showed that the portion of target voiced segments overlapping with masker voiced segments and the similarity between target and masker overlapped voiced segments impacted listeners' speech recognition. Evidence also suggested modulation masking in the spectro-temporal domain interferes with listeners' ability to glimpse the target.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"2960-2973"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142558111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Vahid Delaram, Margaret K Miller, Rohit M Ananthanarayana, Allison Trine, Emily Buss, G Christopher Stecker, Brian B Monson
{"title":"Gender and speech material effects on the long-term average speech spectrum, including at extended high frequencies.","authors":"Vahid Delaram, Margaret K Miller, Rohit M Ananthanarayana, Allison Trine, Emily Buss, G Christopher Stecker, Brian B Monson","doi":"10.1121/10.0034231","DOIUrl":"10.1121/10.0034231","url":null,"abstract":"<p><p>Gender and language effects on the long-term average speech spectrum (LTASS) have been reported, but typically using recordings that were bandlimited and/or failed to accurately capture extended high frequencies (EHFs). Accurate characterization of the full-band LTASS is warranted given recent data on the contribution of EHFs to speech perception. The present study characterized the LTASS for high-fidelity, anechoic recordings of males and females producing Bamford-Kowal-Bench sentences, digits, and unscripted narratives. Gender had an effect on spectral levels at both ends of the spectrum: males had higher levels than females below approximately 160 Hz, owing to lower fundamental frequencies; females had ∼4 dB higher levels at EHFs, but this effect was dependent on speech material. Gender differences were also observed at ∼300 Hz, and between 800 and 1000 Hz, as previously reported. Despite differences in phonetic content, there were only small, gender-dependent differences in EHF levels across speech materials. EHF levels were highly correlated across materials, indicating relative consistency within talkers. Our findings suggest that LTASS levels at EHFs are influenced primarily by talker and gender, highlighting the need for future research to assess whether EHF cues are more audible for female speech than for male speech.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"156 5","pages":"3056-3066"},"PeriodicalIF":2.1,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11540443/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142583198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}