Title :
Cache based recurrent neural network language model inference for first pass speech recognition
Author :
Zhiheng Huang ; Zweig, Geoffrey ; Dumoulin, Benoit
Author_Institution :
Speech at Microsoft, Sunnyvale, CA, USA
Abstract :
Recurrent neural network language models (RNNLMs) have recently produced improvements on language processing tasks ranging from machine translation to word tagging and speech recognition. To date, however, the computational expense of RNNLMs has hampered their application to first pass decoding. In this paper, we show that by restricting the RNNLM calls to those words that receive a reasonable score according to a n-gram model, and by deploying a set of caches, we can reduce the cost of using an RNNLM in the first pass to that of using an additional n-gram model. We compare this scheme to lattice rescoring, and find that they produce comparable results for a Bing Voice search task. The best performance results from rescoring a lattice that is itself created with a RNNLM in the first pass.
Keywords :
language translation; recurrent neural nets; search problems; speech recognition; Bing Voice search task; RNNLM calls; cache based recurrent neural network language model inference; computational expense; first pass speech recognition; language processing tasks; lattice rescoring; machine translation; n-gram model; word tagging; Computational modeling; Data models; Decoding; Hidden Markov models; History; Recurrent neural networks; Speech recognition; cache; computational efficiency; recurrent neural network language model; voice search;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
DOI :
10.1109/ICASSP.2014.6854827