Title :
Unsupervised CV language model adaptation based on direct likelihood maximization sentence selection
Author :
Shinozaki, Takahiro ; Horiuchi, Yasuo ; Kuroiwa, Shingo
Author_Institution :
Div. of Inf. Sci., Chiba Univ., Chiba, Japan
Abstract :
Direct likelihood maximization selection (DLMS) selects a subset of language model training data so that likelihood of in-domain development data is maximized. By using recognition hypothesis instead of the in-domain development data, it can be used for unsupervised adaptation. We apply DLMS to iterative unsupervised adaptation for presentation speech recognition. A problem of the iterative unsupervised adaptation is that adapted models are estimated including recognition errors and it limits the adaptation performance. To solve the problem, we introduce the framework of unsupervised cross-validation (CV) adaptation that has originally been proposed for acoustic model adaptation. Large vocabulary speech recognition experiments show that the CV approach is effective for DLMS based adaptation reducing 19.3% of error rate by an initial model to 18.0%.
Keywords :
iterative methods; maximum likelihood estimation; speech recognition; DLMS; acoustic model adaptation; direct likelihood maximization sentence selection; in-domain development data; iterative unsupervised adaptation; language model training data subset; recognition hypothesis; speech recognition presentation; unsupervised CV language model adaptation; unsupervised cross-validation adaptation; vocabulary speech recognition; Adaptation models; Data models; Error analysis; Hidden Markov models; Speech recognition; Training; Training data; Cross-validation; language model; relative entropy; sentence selection; unsupervised adaptation;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
Conference_Location :
Kyoto
Print_ISBN :
978-1-4673-0045-2
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2012.6289050