Title :
Variational approximation of long-span language models for lvcsr
Author :
Anoop Deoras;Tomáš Mikolov;Stefan Kombrink;Martin Karafiát;Sanjeev Khudanpur
Author_Institution :
HLTCOE and CLSP, Johns Hopkins University, Baltimore MD 21218, USA
fDate :
5/1/2011 12:00:00 AM
Abstract :
Long-span language models that capture syntax and semantics are seldom used in the first pass of large vocabulary continuous speech recognition systems due to the prohibitive search-space of sentence-hypotheses. Instead, an N-best list of hypotheses is created using tractable n-gram models, and rescored using the long-span models. It is shown in this paper that computationally tractable variational approximations of the long-span models are a better choice than standard ra-gram models for first pass decoding. They not only result in a better first pass output, but also produce a lattice with a lower oracle word error rate, and rescoring the N-best list from such lattices with the long-span models requires a smaller N to attain the same accuracy. Empirical results on the WSJ, MIT Lectures, NIST 2007 Meeting Recognition and NIST 2001 Conversational Telephone Recognition data sets are presented to support these claims.
Keywords :
"Decoding","Recurrent neural networks","Lattices","Computational modeling","Approximation methods","Speech recognition","Data models"
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on
Print_ISBN :
978-1-4577-0538-0
Electronic_ISBN :
2379-190X
DOI :
10.1109/ICASSP.2011.5947612