Title :
Low-rank matrix factorization for Deep Neural Network training with high-dimensional output targets
Author :
Sainath, Tara N. ; Kingsbury, Brian ; Sindhwani, Vikas ; Arisoy, Ebru ; Ramabhadran, Bhuvana
Author_Institution :
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
Abstract :
While Deep Neural Networks (DNNs) have achieved tremendous success for large vocabulary continuous speech recognition (LVCSR) tasks, training of these networks is slow. One reason is that DNNs are trained with a large number of training parameters (i.e., 10-50 million). Because networks are trained with a large number of output targets to achieve good performance, the majority of these parameters are in the final weight layer. In this paper, we propose a low-rank matrix factorization of the final weight layer. We apply this low-rank technique to DNNs for both acoustic modeling and language modeling. We show on three different LVCSR tasks ranging between 50-400 hrs, that a low-rank factorization reduces the number of parameters of the network by 30-50%. This results in roughly an equivalent reduction in training time, without a significant loss in final recognition accuracy, compared to a full-rank representation.
Keywords :
matrix decomposition; neural nets; speech recognition; LVCSR; acoustic modeling; deep neural network; high-dimensional output targets; language modeling; large vocabulary continuous speech recognition; low-rank matrix factorization; time 50 hr to 400 hr; Accuracy; Acoustics; Hidden Markov models; Neural networks; Speech; Speech recognition; Training; Deep Neural Networks; Speech Recognition;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
DOI :
10.1109/ICASSP.2013.6638949