DocumentCode
3530387
Title
Using collective information in semi-supervised learning for speech recognition
Author
Varadarajan, Balakrishnan ; Yu, Dong ; Deng, Li ; Acero, Alex
Author_Institution
Johns Hopkins Univ., Baltimore, MD
fYear
2009
fDate
19-24 April 2009
Firstpage
4633
Lastpage
4636
Abstract
Training accurate acoustic models typically requires a large amount of transcribed data, which can be expensive to obtain. In this paper, we describe a novel semi-supervised learning algorithm for automatic speech recognition. The algorithm determines whether a hypothesized transcription should be used in the training by taking into consideration collective information from all utterances available instead of solely based on the confidence from that utterance itself. It estimates the expected entropy reduction each utterance and transcription pair may cause to the whole unlabeled dataset and choose the ones with the positive gains. We compare our algorithm with existing confidence-based semi-supervised learning algorithm and show that the former can consistently outperform the latter when the same amount of utterances is selected into the training set. We also indicate that our algorithm may determine the cutoff-point in a principled way by demonstrating that the point it finds is very close to the achievable peak point.
Keywords
entropy; learning (artificial intelligence); speech recognition; collective information; entropy reduction; hypothesized transcription; semi-supervised learning; speech recognition; Automatic speech recognition; Databases; Entropy; Lattices; Semisupervised learning; Speech recognition; Speech synthesis; Training data; Semi-supervised learning; collective information; confidence; entropy reduction; lattice;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on
Conference_Location
Taipei
ISSN
1520-6149
Print_ISBN
978-1-4244-2353-8
Electronic_ISBN
1520-6149
Type
conf
DOI
10.1109/ICASSP.2009.4960663
Filename
4960663
Link To Document