DocumentCode :
2280263
Title :
Unsupervised training of acoustic models for large vocabulary continuous speech recognition
Author :
Wessel, Frank ; Ney, Hermann
Author_Institution :
Comput. Sci. Dept., RWTH Aachen Univ. of Technol., Germany
fYear :
2001
fDate :
2001
Firstpage :
307
Lastpage :
310
Abstract :
For speech recognition systems, the amount of acoustic training data is of crucial importance. In the past, large amounts of speech were recorded and transcribed manually for training. Since untranscribed speech is available in various forms these days, the unsupervised training of a speech recognizer on recognized transcriptions is studied. A low-cost recognizer trained with only one hour of manually transcribed speech is used to recognize 72 hours of untranscribed acoustic data. These transcriptions are then used in combination with confidence measures to train an improved recognizer. The effect of confidence measures which are used to detect possible recognition errors is studied systematically. Finally, the unsupervised training is applied iteratively. Using this method, the recognizer is trained with very little manual effort while losing only 14.3% relative on the Broadcast News ´96 and 18.6% relative on the Broadcast News ´98 evaluation test sets.
Keywords :
acoustic signal processing; iterative methods; speech recognition; unsupervised learning; acoustic models; confidence measures; continuous speech recognition; recognition errors; unsupervised training; Acoustic signal detection; Broadcasting; Computer science; Error analysis; Linear discriminant analysis; Personnel; Speech recognition; Testing; Training data; Vocabulary;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Automatic Speech Recognition and Understanding, 2001. ASRU '01. IEEE Workshop on
Print_ISBN :
0-7803-7343-X
Type :
conf
DOI :
10.1109/ASRU.2001.1034648
Filename :
1034648
Link To Document :
بازگشت