DocumentCode :
2329497
Title :
Efficient training of large neural networks for language modeling
Author :
Schwenk, Holger
Author_Institution :
LIMSI, CNRS, Orsay, France
Volume :
4
fYear :
2004
fDate :
25-29 July 2004
Firstpage :
3059
Abstract :
Recently there has been increasing interest in using neural networks for language modeling. In contrast to the well-known backoff n-gram language models, the neural network approach tries to limit the data sparseness problem by performing the estimation in a continuous space, allowing by this means smooth interpolations. The complexity to train such a model and to calculate one n-gram probability is however several orders of magnitude higher than for the backoff models, making the new approach difficult to use in real applications. In this paper several techniques are presented that allow the use of a neural network language model in a large vocabulary speech recognition system, in particular very, fast lattice rescoring and efficient training of large neural networks on training corpora of over 10 million words. The described approach achieves significant word error reductions with respect to a carefully tuned 4-gram backoff language model in a state of the art conversational speech recognizer for the DARPA rich transcriptions evaluations.
Keywords :
computational complexity; neural nets; probability; speech recognition; backoff language model; data sparseness problem; language modeling; neural networks; transcriptions evaluations; vocabulary speech recognition system; Context modeling; Electronic mail; Interpolation; Natural languages; Neural networks; Predictive models; Probability; Speech analysis; Speech recognition; Vocabulary;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2004. Proceedings. 2004 IEEE International Joint Conference on
ISSN :
1098-7576
Print_ISBN :
0-7803-8359-1
Type :
conf
DOI :
10.1109/IJCNN.2004.1381158
Filename :
1381158
Link To Document :
بازگشت