DocumentCode
37724
Title
Optimization Techniques to Improve Training Speed of Deep Neural Networks for Large Speech Tasks
Author
Sainath, Tara N. ; Kingsbury, Brian ; Soltau, Hagen ; Ramabhadran, Bhuvana
Author_Institution
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
Volume
21
Issue
11
fYear
2013
fDate
Nov. 2013
Firstpage
2267
Lastpage
2276
Abstract
While Deep Neural Networks (DNNs) have achieved tremendous success for large vocabulary continuous speech recognition (LVCSR) tasks, training these networks is slow. Even to date, the most common approach to train DNNs is via stochastic gradient descent, serially on one machine. Serial training, coupled with the large number of training parameters (i.e., 10-50 million) and speech data set sizes (i.e., 20-100 million training points) makes DNN training very slow for LVCSR tasks. In this work, we explore a variety of different optimization techniques to improve DNN training speed. This includes parallelization of the gradient computation during cross-entropy and sequence training, as well as reducing the number of parameters in the network using a low-rank matrix factorization. Applying the proposed optimization techniques, we show that DNN training can be sped up by a factor of 3 on a 50-hour English Broadcast News (BN) task with no loss in accuracy. Furthermore, using the proposed techniques, we are able to train DNNs on a 300-hr Switchboard (SWB) task and a 400-hr English BN task, showing improvements between 9-30% relative over a state-of-the art GMM/HMM system while the number of parameters of the DNN is smaller than the GMM/HMM system.
Keywords
entropy; hidden Markov models; matrix decomposition; neural nets; optimisation; speech recognition; DNN training; English broadcast news; GMM-HMM system; LVCSR tasks; cross-entropy; deep neural networks; gradient computation; large speech tasks; large vocabulary continuous speech recognition; low-rank matrix factorization; optimization techniques; parallelization; serial training; speech data set sizes; stochastic gradient descent; switchboard task; time 300 hr; time 50 hour; training speed; Hidden Markov models; Large scale systems; Linear programming; Neural networks; Optimization; Pattern recognition; Speech recognition; deep neural networks; parallel optimization techniques;
fLanguage
English
Journal_Title
Audio, Speech, and Language Processing, IEEE Transactions on
Publisher
ieee
ISSN
1558-7916
Type
jour
DOI
10.1109/TASL.2013.2284378
Filename
6619439
Link To Document