DocumentCode :
730715
Title :
Convolutional, Long Short-Term Memory, fully connected Deep Neural Networks
Author :
Sainath, Tara N. ; Vinyals, Oriol ; Senior, Andrew ; Sak, Hasim
Author_Institution :
Google, Inc., New York, NY, USA
fYear :
2015
fDate :
19-24 April 2015
Firstpage :
4580
Lastpage :
4584
Abstract :
Both Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) have shown improvements over Deep Neural Networks (DNNs) across a wide variety of speech recognition tasks. CNNs, LSTMs and DNNs are complementary in their modeling capabilities, as CNNs are good at reducing frequency variations, LSTMs are good at temporal modeling, and DNNs are appropriate for mapping features to a more separable space. In this paper, we take advantage of the complementarity of CNNs, LSTMs and DNNs by combining them into one unified architecture. We explore the proposed architecture, which we call CLDNN, on a variety of large vocabulary tasks, varying from 200 to 2,000 hours. We find that the CLDNN provides a 4-6% relative improvement in WER over an LSTM, the strongest of the three individual models.
Keywords :
neural nets; speech recognition; CNN; DNN; LSTM; convolutional memory; convolutional neural networks; frequency variations; fully connected deep neural networks; long short-term memory; speech recognition; Context; Hidden Markov models; Neural networks; Noise measurement; Speech; Speech recognition; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
Type :
conf
DOI :
10.1109/ICASSP.2015.7178838
Filename :
7178838
Link To Document :
بازگشت