Title :
Full HMM Training for Minimizing Generation Error in Synthesis
Author :
Yi-Jian Wu ; Ren-Hua Wang ; Soong, Frank
Author_Institution :
Microsoft Res. Asia, Beijing, China
Abstract :
In maximum-likelihood (ML) based HMM synthesis, the generated trajectory of a sentence in the training set is in general does not reproduce the trajectory of the original one. To overcome this shortcoming, a minimum generation error (MGE) criterion has been previously proposed. In this paper, a complete MGE-based HMM training is introduced, where the MGE criterion is applied to the entire training process, including context-dependent HMM training, context-dependent HMM clustering and clustered HMM training. In this procedure, the HMMs are trained to minimize the generation error of training data, which is in line with the HMM-based synthesis. From the experiments, the quality of synthesized speech is improved after applying the MGE criterion to the whole training process.
Keywords :
hidden Markov models; learning (artificial intelligence); maximum likelihood estimation; speech synthesis; clustered HMM training; context-dependent HMM clustering; context-dependent HMM training; maximum-likelihood based HMM synthesis; minimum generation error criterion; synthesized speech; Asia; Computational efficiency; Hidden Markov models; Optical wavelength conversion; Speech processing; Speech synthesis; Training data; HMM; maximum likelihood; minimum generation error; speech synthesis;
Conference_Titel :
Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference on
Conference_Location :
Honolulu, HI
Print_ISBN :
1-4244-0727-3
DOI :
10.1109/ICASSP.2007.366963