DocumentCode :
1693369
Title :
Learning discriminative basis coefficients for eigenspace MLLR unsupervised adaptation
Author :
Yajie Miao ; Metze, Florian ; Waibel, Alex
Author_Institution :
Language Technol. Inst., Carnegie Mellon Univ., Pittsburgh, PA, USA
fYear :
2013
Firstpage :
7927
Lastpage :
7931
Abstract :
Eigenspace MLLR is effective for fast adaptation when the amount of adaptation data is limited, e.g., less than 5s. The general motivation is to represent the MLLR transform as a linear combination of basis matrices. In this paper, we present a framework to estimate a speaker-independent discriminative transform over the combination coefficients. This discriminative basis coefficients transform (DBCT) is learned by optimizing discriminative criteria over all the training speakers. During recognition, the ML basis coefficients for each testing speaker are firstly found, on which DBCT is applied to give the final MLLR transform discrimination ability. Experiments show that DBCT results in consistent WER reduction in unsupervised adaptation, compared with both standard ML and discriminatively trained transforms.
Keywords :
maximum likelihood estimation; regression analysis; speaker recognition; unsupervised learning; DBCT; MLLR transform; WER reduction; basis matrices; combination coefficients; discriminative basis coefficients transform; eigenspace MLLR; learning discriminative basis coefficients; linear combination; maximum likelihood linear regression; speaker recognition; speaker-independent discriminative transform; unsupervised adaptation; Adaptation models; Estimation; Hidden Markov models; Speech; Testing; Training; Transforms; Speaker adaptation; discriminative training; speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
ISSN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2013.6639208
Filename :
6639208
Link To Document :
بازگشت