{"title":"利用时间自适应窗提高音频空间化的时频稀疏性","authors":"P. Gaddipati, N. Dave, P. Rao, R. Velmurugan","doi":"10.1109/NCC.2010.5430178","DOIUrl":null,"url":null,"abstract":"We propose a new time-adaptive windowing technique to obtain a sparse time-frequency representation for audio signals. This transformation helps in providing better source separation from stereo mixtures for improved subsequent spatial rendering over headphones. We start with standard stereo audio recordings, transform them to a sparse representation and then estimate the mixing parameters to be used for source separation. The performance of the new representation is compared with existing methods via the accuracy of mixing parameters estimation for a test dataset of multi-speaker stereo mixtures.","PeriodicalId":130953,"journal":{"name":"2010 National Conference On Communications (NCC)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving time-frequency sparsity for audio spatialization by time-adaptive windowing\",\"authors\":\"P. Gaddipati, N. Dave, P. Rao, R. Velmurugan\",\"doi\":\"10.1109/NCC.2010.5430178\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We propose a new time-adaptive windowing technique to obtain a sparse time-frequency representation for audio signals. This transformation helps in providing better source separation from stereo mixtures for improved subsequent spatial rendering over headphones. We start with standard stereo audio recordings, transform them to a sparse representation and then estimate the mixing parameters to be used for source separation. The performance of the new representation is compared with existing methods via the accuracy of mixing parameters estimation for a test dataset of multi-speaker stereo mixtures.\",\"PeriodicalId\":130953,\"journal\":{\"name\":\"2010 National Conference On Communications (NCC)\",\"volume\":\"33 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-03-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 National Conference On Communications (NCC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NCC.2010.5430178\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 National Conference On Communications (NCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NCC.2010.5430178","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Improving time-frequency sparsity for audio spatialization by time-adaptive windowing
We propose a new time-adaptive windowing technique to obtain a sparse time-frequency representation for audio signals. This transformation helps in providing better source separation from stereo mixtures for improved subsequent spatial rendering over headphones. We start with standard stereo audio recordings, transform them to a sparse representation and then estimate the mixing parameters to be used for source separation. The performance of the new representation is compared with existing methods via the accuracy of mixing parameters estimation for a test dataset of multi-speaker stereo mixtures.