{"title":"Computational cost reduction of long short-term memory based on simultaneous compression of input and hidden state","authors":"T. Masuko","doi":"10.1109/ASRU.2017.8268926","DOIUrl":null,"url":null,"abstract":"Long short-term memory (LSTM) has been successfully applied to acoustic modeling for automatic speech recognition (ASR). However, because of its complicated structure, LSTM requires high computational cost especially when the number of dimensions of memory cell is sufficiently high to get good ASR performance. In this paper, we present a novel technique to reduce computational cost of LSTM in which the input and the previous hidden state vectors are simultaneously compressed with a linear projection layer. From experimental results, it is shown that the proposed technique outperforms a standard LSTM and an LSTM with a recurrent projection layer. It is also shown that in the proposed technique ASR performance is improved by increasing the number of dimensions of memory cell when the sizes of models are comparable.","PeriodicalId":290868,"journal":{"name":"2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ASRU.2017.8268926","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8
Abstract
Long short-term memory (LSTM) has been successfully applied to acoustic modeling for automatic speech recognition (ASR). However, because of its complicated structure, LSTM requires high computational cost especially when the number of dimensions of memory cell is sufficiently high to get good ASR performance. In this paper, we present a novel technique to reduce computational cost of LSTM in which the input and the previous hidden state vectors are simultaneously compressed with a linear projection layer. From experimental results, it is shown that the proposed technique outperforms a standard LSTM and an LSTM with a recurrent projection layer. It is also shown that in the proposed technique ASR performance is improved by increasing the number of dimensions of memory cell when the sizes of models are comparable.