Title :
Neuron sparseness versus connection sparseness in deep neural network for large vocabulary speech recognition
Author :
Jian Kang ; Cheng Lu ; Meng Cai ; Wei-Qiang Zhang ; Jia Liu
Author_Institution :
Dept. of Electron. Eng., Tsinghua Univ., Beijing, China
Abstract :
Exploiting sparseness in deep neural networks is an important method for reducing the computational cost. In this paper, we study neuron sparseness in deep neural networks for acoustic modeling. For the feed-forward stage, we only activate neurons whose input values are larger than a given threshold, and set the outputs of inactive nodes to zero. Thus, only a few nonzero outputs are fed to the next layer. Using this method, the output vector of each hidden layer becomes very sparse, so that the computational cost of the feed-forward algorithm can be reduced by adopting sparse matrix operations. The proposed method is evaluated in both small and large vocabulary speech recognition tasks, and results demonstrate that we can reduce the nonzero outputs to fewer than 20% of the total number of hidden nodes, without sacrificing speech recognition performance.
Keywords :
feedforward; neural nets; sparse matrices; speech recognition; vocabulary; acoustic modeling; connection sparseness; deep neural network; feed-forward algorithm; large vocabulary speech recognition; neuron sparseness; sparse matrix operations; Artificial neural networks; Complexity theory; Context; Hidden Markov models; Neurons; Vocabulary; acoustic modeling; deep neural network; sparseness; speech recognition;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
DOI :
10.1109/ICASSP.2015.7178913