{"title":"A Multimode Multi-Band Excitation with Linear Prediction vocoder","authors":"Yanxia Liang, Jiawei Yang, Ye Li","doi":"10.1109/WCINS.2010.5541932","DOIUrl":null,"url":null,"abstract":"This paper presents a Multimode Multi-Band Excitation with Linear Prediction model (MMBE-LP) at 2.35kbps. Unvoiced/Voiced (U/V) decisions and spectrum amplitudes estimations are improved in this vocoder compared with MBE vocoder. For better quantization results, different codebooks are used for different modes of U/V decision, so the number of sub-bands in a frame is fixed. Spectral amplitudes are estimated by Linear Prediction Mode and quantized by MSVQ (Multi-Stage Vector Quantization). Simulation results show that the unvoiced and voiced parts of synthetic speech are coherent with the parts of original speech obviously.","PeriodicalId":156036,"journal":{"name":"2010 IEEE International Conference on Wireless Communications, Networking and Information Security","volume":"80 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 IEEE International Conference on Wireless Communications, Networking and Information Security","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WCINS.2010.5541932","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
This paper presents a Multimode Multi-Band Excitation with Linear Prediction model (MMBE-LP) at 2.35kbps. Unvoiced/Voiced (U/V) decisions and spectrum amplitudes estimations are improved in this vocoder compared with MBE vocoder. For better quantization results, different codebooks are used for different modes of U/V decision, so the number of sub-bands in a frame is fixed. Spectral amplitudes are estimated by Linear Prediction Mode and quantized by MSVQ (Multi-Stage Vector Quantization). Simulation results show that the unvoiced and voiced parts of synthetic speech are coherent with the parts of original speech obviously.