DocumentCode :
1361792
Title :
Introduction to the Special Section on Deep Learning for Speech and Language Processing
Author :
Dong Yu ; Hinton, Geoffrey ; Morgan, Nigel ; Jen-Tzung Chien ; Sagayama, Shigeki
Author_Institution :
Microsoft Res., Redmond, WA, USA
Volume :
20
Issue :
1
fYear :
2012
Firstpage :
4
Lastpage :
6
Abstract :
Current speech recognition systems, for example, typically use Gaussian mixture models (GMMs), to estimate the observation (or emission) probabilities of hidden Markov models (HMMs), and GMMs are generative models that have only one layer of latent variables. Instead of developing more powerful models, most of the research effort has gone into finding better ways of estimating the GMM parameters so that error rates are decreased or the margin between different classes is increased. The same observation holds for natural language processing (NLP) in which maximum entropy (MaxEnt) models and conditional random fields (CRFs) have been popular for the last decade. Both of these approaches use shallow models whose success largely depends on the use of carefully handcrafted features.
Keywords :
error statistics; learning (artificial intelligence); maximum entropy methods; natural language processing; speech recognition; GMM parameter; Gaussian mixture model; MaxEnt model; conditional random field; hidden Markov model; maximum entropy (MaxEnt) model; natural language processing; speech processing; speech recognition system; Automatic speech recognition; Hidden Markov models; Machine learning; Special issues and sections; Speech recognition;
fLanguage :
English
Journal_Title :
Audio, Speech, and Language Processing, IEEE Transactions on
Publisher :
ieee
ISSN :
1558-7916
Type :
jour
DOI :
10.1109/TASL.2011.2173371
Filename :
6060895
Link To Document :
بازگشت