Title :
Understanding how Deep Belief Networks perform acoustic modelling
Author :
Mohamed, Abdel-rahman ; Hinton, Geoffrey ; Penn, Gerald
Author_Institution :
Dept. of Comput. Sci., Univ. of Toronto, Toronto, ON, Canada
Abstract :
Deep Belief Networks (DBNs) are a very competitive alternative to Gaussian mixture models for relating states of a hidden Markov model to frames of coefficients derived from the acoustic input. They are competitive for three reasons: DBNs can be fine-tuned as neural networks; DBNs have many non-linear hidden layers; and DBNs are generatively pre-trained. This paper illustrates how each of these three aspects contributes to the DBN´s good recognition performance using both phone recognition performance on the TIMIT corpus and a dimensionally reduced visualization of the relationships between the feature vectors learned by the DBNs that preserves the similarity structure of the feature vectors at multiple scales. The same two methods are also used to investigate the most suitable type of input representation for a DBN.
Keywords :
Gaussian processes; acoustic signal processing; belief networks; hidden Markov models; neural nets; speech recognition; vectors; DBN; Gaussian mixture models; TIMIT corpus; acoustic modelling; deep belief networks; dimensionally reduced visualization; feature vectors; hidden Markov model; neural networks; Biological neural networks; Hidden Markov models; Mel frequency cepstral coefficient; Speech; Training; Vectors; Deep belief networks; acoustic modeling; neural networks;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
Conference_Location :
Kyoto
Print_ISBN :
978-1-4673-0045-2
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2012.6288863