{"title":"A neural network-based text independent voice recognition system","authors":"K. Kuah, M. Bodruzzaman, S. Zein-Sabatto","doi":"10.1109/SECON.1994.324282","DOIUrl":null,"url":null,"abstract":"A text-independent voice recognition experiment was conducted using an artificial neural network. The speech data were collected from three different speakers uttering thirteen different words. Each word was repeated ten times. The speech data were then pre-processed for signal conditioning. A total of 12 feature parameters were obtained from Cepstral coefficients via a linear predictive coding (LPC). These feature parameters then served as inputs to the neural network for speaker classification. A standard two-layer feedforward neural network was trained to identify different feature sets associated with the corresponding speakers. The network was tested for the remaining unseen words in text-independent mode. The results were very promising with a voice recognition accuracy of more than 90%. The success rate could be increased by adding more utterances from each speaker.<<ETX>>","PeriodicalId":119615,"journal":{"name":"Proceedings of SOUTHEASTCON '94","volume":"105 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1994-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of SOUTHEASTCON '94","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SECON.1994.324282","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
A text-independent voice recognition experiment was conducted using an artificial neural network. The speech data were collected from three different speakers uttering thirteen different words. Each word was repeated ten times. The speech data were then pre-processed for signal conditioning. A total of 12 feature parameters were obtained from Cepstral coefficients via a linear predictive coding (LPC). These feature parameters then served as inputs to the neural network for speaker classification. A standard two-layer feedforward neural network was trained to identify different feature sets associated with the corresponding speakers. The network was tested for the remaining unseen words in text-independent mode. The results were very promising with a voice recognition accuracy of more than 90%. The success rate could be increased by adding more utterances from each speaker.<>