DocumentCode :
3162854
Title :
Understanding how Deep Belief Networks perform acoustic modelling
Author :
Mohamed, Abdel-rahman ; Hinton, Geoffrey ; Penn, Gerald
Author_Institution :
Dept. of Comput. Sci., Univ. of Toronto, Toronto, ON, Canada
fYear :
2012
fDate :
25-30 March 2012
Firstpage :
4273
Lastpage :
4276
Abstract :
Deep Belief Networks (DBNs) are a very competitive alternative to Gaussian mixture models for relating states of a hidden Markov model to frames of coefficients derived from the acoustic input. They are competitive for three reasons: DBNs can be fine-tuned as neural networks; DBNs have many non-linear hidden layers; and DBNs are generatively pre-trained. This paper illustrates how each of these three aspects contributes to the DBN´s good recognition performance using both phone recognition performance on the TIMIT corpus and a dimensionally reduced visualization of the relationships between the feature vectors learned by the DBNs that preserves the similarity structure of the feature vectors at multiple scales. The same two methods are also used to investigate the most suitable type of input representation for a DBN.
Keywords :
Gaussian processes; acoustic signal processing; belief networks; hidden Markov models; neural nets; speech recognition; vectors; DBN; Gaussian mixture models; TIMIT corpus; acoustic modelling; deep belief networks; dimensionally reduced visualization; feature vectors; hidden Markov model; neural networks; Biological neural networks; Hidden Markov models; Mel frequency cepstral coefficient; Speech; Training; Vectors; Deep belief networks; acoustic modeling; neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
Conference_Location :
Kyoto
ISSN :
1520-6149
Print_ISBN :
978-1-4673-0045-2
Electronic_ISBN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2012.6288863
Filename :
6288863
Link To Document :
بازگشت