Title :
Recurrent neural network language model training with noise contrastive estimation for speech recognition
Author :
Chen, X. ; Liu, X. ; Gales, M.J.F. ; Woodland, P.C.
Author_Institution :
Eng. Dept., Univ. of Cambridge, Cambridge, UK
Abstract :
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to a range of tasks including speech recognition. However, an important issue that limits the quantity of data used, and their possible application areas, is the computational cost in training. A signi??cant part of this cost is associated with the softmax function at the output layer, as this requires a normalization term to be explicitly calculated. This impacts both the training and testing speed, especially when a large output vocabulary is used. To address this problem, noise contrastive estimation (NCE) is explored in RNNLM training. NCE does not require the above normalization during both training and testing. It is insensitive to the output layer size. On a large vocabulary conversational telephone speech recognition task, a doubling in training speed on a GPU and a 56 times speed up in test time evaluation on a CPU were obtained.
Keywords :
graphics processing units; language translation; learning (artificial intelligence); recurrent neural nets; speech recognition; vocabulary; CPU; GPU; NCE; RNNLM training; computational cost; large vocabulary conversational telephone speech recognition task; noise contrastive estimation; normalization term; recurrent neural network language model training; softmax function; History; Noise; Recurrent neural networks; Speech recognition; Testing; Training; Vocabulary; GPU; language model; noise contrastive estimation; recurrent neural network; speech recognition;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
DOI :
10.1109/ICASSP.2015.7179005