DocumentCode
79502
Title
Efficient One-Pass Decoding with NNLM for Speech Recognition
Author
Yongzhe Shi ; Wei-Qiang Zhang ; Meng Cai ; Jia Liu
Author_Institution
Dept. of Electron. & Eng., Tsinghua Univ., Beijing, China
Volume
21
Issue
4
fYear
2014
fDate
Apr-14
Firstpage
377
Lastpage
381
Abstract
Neural network language model (NNLM) has achieved very good results in the field of speech recognition, machine translation, etc. Direct decoding with NNLM is challenging for the overwhelmingly heavy burden in complexity. Most of the previous work focused on rescoring the N-best list and lattice with NNLM in the second pass. In this work, several techniques are explored to directly incorporate the NNLM into the decoder of speech recognition. A novel training algorithm based on variance regularization is proposed to approximate the softmax-normalizing factor as a constant for fast evaluation. Also, the evaluation of NNLM is further speeded up via our advanced storage. Moreover, a simple cache-based strategy is explored to avoid redundant computations during the decoding process. To the authors´ knowledge, it is the first time to directly incorporate NNLM into decoding. We evaluate our proposed methods on an English-Switchboard phone-call speech-to-text task. Experimental results show that incorporating the NNLM into the decoder significantly reduces the word error rate (WER) by 1.5% and 1.4% absolutely on the Hub5´00-SWB and RT03S-FSH sets, respectively. Also, the decoding with NNLM is twice as fast as the baseline at the same word error rate.
Keywords
decoding; neural nets; speech recognition; NNLM; cache based strategy; direct decoding; machine translation; neural network language model; one pass decoding; softmax normalizing factor; speech recognition; variance regularization; word error rate; Approximation methods; Artificial neural networks; Complexity theory; Decoding; Lattices; Speech recognition; Neural network language model; one-pass decoding; speech recognition;
fLanguage
English
Journal_Title
Signal Processing Letters, IEEE
Publisher
ieee
ISSN
1070-9908
Type
jour
DOI
10.1109/LSP.2014.2303136
Filename
6727395
Link To Document