Title : 
Improving the training and evaluation efficiency of recurrent neural network language models
         
        
            Author : 
Chen, X. ; Liu, X. ; Gales, M.J.F. ; Woodland, P.C.
         
        
            Author_Institution : 
Eng. Dept., Univ. of Cambridge, Cambridge, UK
         
        
        
        
        
        
            Abstract : 
Recurrent neural network language models (RNNLMs) are becoming increasingly popular for speech recognition. Previously, we have shown that RNNLMs with a full (non-classed) output layer (F-RNNLMs) can be trained efficiently using a GPU giving a large reduction in training time over conventional class-based models (C-RNNLMs) on a standard CPU. However, since test-time RNNLM evaluation is often performed entirely on a CPU, standard F-RNNLMs are inefficient since the entire output layer needs to be calculated for normalisation. In this paper, it is demonstrated that C-RNNLMs can be efficiently trained on a GPU, using our spliced sentence bunch technique which allows good CPU test-time performance (42× speedup over F-RNNLM). Furthermore, the performance of different classing approaches is investigated. We also examine the use of variance regularisation of the softmax denominator for F-RNNLMs and show that it allows F-RNNLMs to be efficiently used in test (56× speedup on a CPU). Finally the use of two GPUs for F-RNNLM training using pipelining is described and shown to give a reduction in training time over a single GPU by a factor of 1.6×.
         
        
            Keywords : 
graphics processing units; natural language processing; pipeline processing; recurrent neural nets; speech recognition; C-RNNLM; CPU; F-RNNLM; GPU; class-based model; pipelining; recurrent neural network language model; softmax denominator; speech recognition; spliced sentence bunch technique; test-time RNNLM evaluation; training time reduction; variance regularisation; Boolean functions; Data structures; Indexes; Pipeline processing; Recurrent neural networks; GPU; language models; recurrent neural network; speech recognition;
         
        
        
        
            Conference_Titel : 
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
         
        
            Conference_Location : 
South Brisbane, QLD
         
        
        
            DOI : 
10.1109/ICASSP.2015.7179003