Title :
Efficient training algorithms for HMMs using incremental estimation
Author :
Gotoh, Yoshihiko ; Hochberg, Michael M. ; Silverman, Harvey F.
Author_Institution :
Dept. of Comput. Sci., Sheffield Univ., UK
fDate :
11/1/1998 12:00:00 AM
Abstract :
Typically, parameter estimation for a hidden Markov model (HMM) is performed using an expectation-maximization (EM) algorithm with the maximum-likelihood (ML) criterion. The EM algorithm is an iterative scheme that is well-defined and numerically stable, but convergence may require a large number of iterations. For speech recognition systems utilizing large amounts of training material, this results in long training times. This paper presents an incremental estimation approach to speed-up the training of HMMs without any loss of recognition performance. The algorithm selects a subset of data from the training set, updates the model parameters based on the subset, and then iterates the process until convergence of the parameters. The advantage of this approach is a substantial increase in the number of iterations of the EM algorithm per training token, which leads to faster training. In order to achieve reliable estimation from a small fraction of the complete data set at each iteration, two training criteria are studied; ML and maximum a posteriori (MAP) estimation. Experimental results show that the training of the incremental algorithms is substantially faster than the conventional (batch) method and suffers no loss of recognition performance. Furthermore, the incremental MAP based training algorithm improves performance over the batch version
Keywords :
convergence of numerical methods; hidden Markov models; iterative methods; maximum likelihood estimation; speech recognition; EM algorithm; HMM; MAP estimation; MLE; convergence; efficient training algorithms; expectation-maximization algorithm; experimental results; hidden Markov model; incremental estimation; iterative scheme; maximum a posteriori estimation; maximum-likelihood estimation; model parameters; numerically stable method; parameter estimation; recognition performance; speech recognition systems; training material; training times; Convergence of numerical methods; Helium; Hidden Markov models; Iterative algorithms; Maximum likelihood estimation; Parameter estimation; Performance loss; Speech processing; Speech recognition; Training data;
Journal_Title :
Speech and Audio Processing, IEEE Transactions on