Van Ha Tang, A. Bouzerdoum, S. L. Phung, F. Tivive
{"title":"Multi-view indoor scene reconstruction from compressed through-wall radar measurements using a joint bayesian sparse representation","authors":"Van Ha Tang, A. Bouzerdoum, S. L. Phung, F. Tivive","doi":"10.1109/ICASSP.2015.7178405","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178405","url":null,"abstract":"This paper addresses the problem of scene reconstruction, incorporating wall-clutter mitigation, for compressed multi-view through-the-wall radar imaging. We consider the problem where the scene is sensed using different reduced sets of frequencies at different antennas. A joint Bayesian sparse recovery framework is first employed to estimate the antenna signal coefficients simultaneously, by exploiting the sparsity and correlations between antenna signals. Following joint signal coefficient estimation, a subspace projection technique is applied to segregate the target coefficients from the wall contributions. Furthermore, a multitask linear model is developed to relate the target coefficients to the scene, and a composite scene image is reconstructed by a joint Bayesian sparse framework, taking into account the inter-view dependencies. Experimental results show that the proposed approach improves reconstruction accuracy and produces a composite scene image in which the targets are enhanced and the background clutter is attenuated.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129945246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Supervised sparse coding with local geometrical constraints","authors":"Hanchao Zhang, Jinhua Xu","doi":"10.1109/ICASSP.2015.7178362","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178362","url":null,"abstract":"Sparse coding algorithms with geometrical constraints have received much attention recently. However, these methods are unsupervised and might lead to less discriminative representations. In this paper, we propose a supervised locality-constrained sparse coding method for classification. Two graphs are constructed, a labeled graph and an unlabeled graph. Sparse codes with a labeled geometrical constraint will be more discriminative, however we cannot embed test samples with unknown label into a labeled graph. By coupling the two graphs, we aim to make the difference between sparse codes with labeled and unlabeled geometrical constraints as small as possible. As a result, sparse codes of test data can be obtained with the unlabeled geometrical constraint and the discrimination of the labeled geometrical constraint is maintained. Experiments on some benchmark datasets demonstrate the effectiveness of the proposed method.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128696536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ivan Himawan, P. Motlícek, David Imseng, B. Potard, Namhoon Kim, Jaewon Lee
{"title":"Learning feature mapping using deep neural network bottleneck features for distant large vocabulary speech recognition","authors":"Ivan Himawan, P. Motlícek, David Imseng, B. Potard, Namhoon Kim, Jaewon Lee","doi":"10.1109/ICASSP.2015.7178830","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178830","url":null,"abstract":"Automatic speech recognition from distant microphones is a difficult task because recordings are affected by reverberation and background noise. First, the application of the deep neural network (DNN)/hidden Markov model (HMM) hybrid acoustic models for distant speech recognition task using AMI meeting corpus is investigated. This paper then proposes a feature transformation for removing reverberation and background noise artefacts from bottleneck features using DNN trained to learn the mapping between distant-talking speech features and close-talking speech bottleneck features. Experimental results on AMI meeting corpus reveal that the mismatch between close-talking and distant-talking conditions is largely reduced, with about 16% relative improvement over conventional bottleneck system (trained on close-talking speech). If the feature mapping is applied to close-talking speech, a minor degradation of 4% relative is observed.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128580185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Base Station clustering in heterogeneous network with finite backhaul capacity","authors":"Qian Zhang, Chen He, Ling-ge Jiang","doi":"10.1109/ICASSP.2015.7178501","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178501","url":null,"abstract":"This paper considers the Base Station (BS) clustering in a downlink heterogeneous network with finite backhaul capacity. We consider a tree structure network where each BS has only one incoming link and several outgoing links. The objective is to maximize the minimum rate among all users while satisfying the backhaul capacity constraint and the per-BS power constraint. We propose an algorithm that combines the bisection and the Alternating Direction Method of Multipliers (ADMM). The bisection search is conducted for the minimum rate. When it is given, we use ADMM to check the feasibility of the network while obeying the backhaul and power constraints. There are two steps involved in ADMM: i) a second order conic programming is used to calculate the beamformer; ii) a closed-form rule is used to determine the BS clustering. Due to the non-convexity and the non-smoothness, ADMM is not guaranteed to solve the problem. Therefore, we propose a revise step to further improve the performance. The simulation results show that the proposed algorithm outperforms the heuristic method and the revise step does improve the performance in all considered scenarios.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128641933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust beamformer and artificial noises for MISO wiretap channels with multiple eavesdroppers","authors":"S. Ohno, Y. Wakasa","doi":"10.1109/ICASSP.2015.7178544","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178544","url":null,"abstract":"We consider MISO wiretap channels with multiple eavesdroppers. Under the deterministic channel uncertainties, the beamformer and the covariance of the artificial noise are jointly designed to minimize the transmit power subject to SINR constraints. Our design problems are resolved by using a semidefinite program, which can be numerically solved. Simulation results are provided to see the effects of channel uncertainties on the transmit power.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128660323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cepstral noise subtraction for robust automatic speech recognition","authors":"R. Rehr, Timo Gerkmann","doi":"10.1109/ICASSP.2015.7177994","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7177994","url":null,"abstract":"The robustness of speech recognizers towards noise can be increased by normalizing the statistical moments of the Mel-frequency cepstral coefficients (MFCCs), e. g. by using cepstral mean normalization (CMN) or cepstral mean and variance normalization (CMVN). The necessary statistics are estimated over a long time window and often, a complete utterance is chosen. Consequently, changes in the background noise can only be tracked to a limited extent which poses a restriction to the performance gain that can be achieved by these techniques. In contrast, algorithms recently developed for single-channel speech enhancement allow to track the background noise quickly. In this paper, we aim at combining speech enhancement techniques and feature normalization methods. For this, we propose to transform an estimate of the noise power spectral density to the MFCC domain, where we subtract it from the noisy MFCCs. This is followed by a conventional CMVN. For background noises that are too instationary for CMVN but can be tracked by the noise estimator, we show that this processing leads to an improvement in comparison to the sole application of CMVN. The observed performance gain emerges especially in low signal-to-noise-ratios.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128972094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rohit Prabhavalkar, R. Álvarez, Carolina Parada, Preetum Nakkiran, Tara N. Sainath
{"title":"Automatic gain control and multi-style training for robust small-footprint keyword spotting with deep neural networks","authors":"Rohit Prabhavalkar, R. Álvarez, Carolina Parada, Preetum Nakkiran, Tara N. Sainath","doi":"10.1109/ICASSP.2015.7178863","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178863","url":null,"abstract":"We explore techniques to improve the robustness of small-footprint keyword spotting models based on deep neural networks (DNNs) in the presence of background noise and in far-field conditions. We find that system performance can be improved significantly, with relative improvements up to 75% in far-field conditions, by employing a combination of multi-style training and a proposed novel formulation of automatic gain control (AGC) that estimates the levels of both speech and background noise. Further, we find that these techniques allow us to achieve competitive performance, even when applied to DNNs with an order of magnitude fewer parameters than our base-line.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129351460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Stability analysis of the FBANC system having delay error in the estimated secondary path model","authors":"Seong-Pil Moon, K. Son, T. Chang","doi":"10.1109/ICASSP.2015.7178053","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178053","url":null,"abstract":"The Feedback active noise control (FBANC) scheme is widely used in portable ANC applications. But the FBANC has un-stability problem caused by the modeling error of the electro-acoustic path in its feedback mechanism. To analyze the stability problem, we propose a new stability analysis method utilizing the magnitude component of the open loop frequency response of the FBANC. With the proposed method, a stability bound equation is obtained in terms of the length of delay error of secondary path, the ANC filter length and the center frequency of primary noise. The stability bounds of the proposed method are verified by comparing with both the original Nyquist condition and the simulation results.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124638846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Interference statistics in a random mmWave ad hoc network","authors":"Andrew Thornburg, T. Bai, R. Heath","doi":"10.1109/ICASSP.2015.7178502","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7178502","url":null,"abstract":"Wireless communication at millimeter wave (mmWave) frequencies is attractive for cellular, local area, and ad hoc networks due to the potential for channels with large bandwidths. As a byproduct of directional beamforming and propagation differences, some studies have claimed that mmWave networks will be noise rather than interference limited. This paper presents a derivation of the instantaneous interference-to-noise ratio (INR) distribution of a mmWave ad hoc network. Random network model of transmitters represented by a Poisson point process with a narrowband channel model is used to derive an approximation of the INR distribution. The analysis shows that the shape of the INR distribution is determined largely by the line-of-sight interferers, which depends on the overall network density and building blockage. A main conclusion drawn is that even with highly directional beamforming, interference can only sometimes be neglected in an ad hoc network. With a reasonable choice of system parameters, the interference is nearly always stronger than the noise power in dense networks.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124766272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dynamic zero-point attracting projection for time-varying sparse signal recovery","authors":"Jiawei Zhou, Laming Chen, Yuantao Gu","doi":"10.1109/ICASSP.2015.7179020","DOIUrl":"https://doi.org/10.1109/ICASSP.2015.7179020","url":null,"abstract":"Sparse signal recovery in the static case has been well studied under the framework of Compressive Sensing (CS), while in recent years more attention has also been paid to the dynamic case. In this paper, enlightened by the idea of modified-CS with partially known support, and based on a non-convex optimization approach, we propose the dynamic zero-point attracting projection (DZAP) algorithm to efficiently recover the slowly time-varying sparse signals. Benefiting from the temporal correlation within signal structures, plus an effective prediction method of the future signal based on previous recoveries incorporated, DZAP achieves high-precision recovery with less measurements or larger sparsity level, which is demonstrated by simulations on both synthetic and real data, accompanied by the comparison with other state-of-the-art reference algorithms.","PeriodicalId":117666,"journal":{"name":"2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129424072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}