{"title":"基于线性预测的深度卷积神经网络去混响语音识别","authors":"Sunchan Park, Yongwon Jeong, M. Kim, H. S. Kim","doi":"10.23919/ELINFOCOM.2018.8330593","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks (CNNs) have been shown to improve classification tasks such as automatic speech recognition (ASR). Furthermore, the CNN with very deep architecture lowered the word error rate (WER) in reverberant and noisy environments. However, DNN-based ASR systems still perform poorly in unseen reverberant conditions. In this paper, we use the weighted prediction error (WPE)-based preprocessing for dereverberation. In our experiments on the ASR task of the REVERB Challenge 2014, the WPE-based processing with eight channels reduced the WER by 20% for the real-condition data using CNN acoustic models with 10 layers.","PeriodicalId":413646,"journal":{"name":"2018 International Conference on Electronics, Information, and Communication (ICEIC)","volume":"177 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Linear prediction-based dereverberation with very deep convolutional neural networks for reverberant speech recognition\",\"authors\":\"Sunchan Park, Yongwon Jeong, M. Kim, H. S. Kim\",\"doi\":\"10.23919/ELINFOCOM.2018.8330593\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Convolutional neural networks (CNNs) have been shown to improve classification tasks such as automatic speech recognition (ASR). Furthermore, the CNN with very deep architecture lowered the word error rate (WER) in reverberant and noisy environments. However, DNN-based ASR systems still perform poorly in unseen reverberant conditions. In this paper, we use the weighted prediction error (WPE)-based preprocessing for dereverberation. In our experiments on the ASR task of the REVERB Challenge 2014, the WPE-based processing with eight channels reduced the WER by 20% for the real-condition data using CNN acoustic models with 10 layers.\",\"PeriodicalId\":413646,\"journal\":{\"name\":\"2018 International Conference on Electronics, Information, and Communication (ICEIC)\",\"volume\":\"177 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 International Conference on Electronics, Information, and Communication (ICEIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/ELINFOCOM.2018.8330593\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 International Conference on Electronics, Information, and Communication (ICEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/ELINFOCOM.2018.8330593","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Linear prediction-based dereverberation with very deep convolutional neural networks for reverberant speech recognition
Convolutional neural networks (CNNs) have been shown to improve classification tasks such as automatic speech recognition (ASR). Furthermore, the CNN with very deep architecture lowered the word error rate (WER) in reverberant and noisy environments. However, DNN-based ASR systems still perform poorly in unseen reverberant conditions. In this paper, we use the weighted prediction error (WPE)-based preprocessing for dereverberation. In our experiments on the ASR task of the REVERB Challenge 2014, the WPE-based processing with eight channels reduced the WER by 20% for the real-condition data using CNN acoustic models with 10 layers.