DocumentCode
417246
Title
Improving broadcast news transcription by lightly supervised discriminative training
Author
Chan, H.Y. ; Woodland, P.C.
Author_Institution
Dept. of Eng., Cambridge Univ., UK
Volume
1
fYear
2004
fDate
17-21 May 2004
Abstract
We present our experiments on lightly supervised discriminative training with large amounts of broadcast news data for which only closed caption transcriptions are available (TDT data). In particular, we use language models biased to the closed-caption transcripts to recognise the audio data, and the recognised transcripts are then used as the training transcriptions for acoustic model training. A range of experiments that use maximum likelihood (ML) training as well as discriminative training based on either maximum mutual information (MMI) or minimum phone error (MPE) are presented. In a 5xRT broadcast news transcription system that includes adaptation, it is shown that reductions in word error rate (WER) in the range of 1% absolute can be achieved. Finally, some experiments on training data selection are presented to compare different methods of "filtering" the transcripts.
Keywords
acoustic signal processing; error statistics; learning (artificial intelligence); speech recognition; LVCSR; ML training; acoustic model training; audio data recognition; broadcast news transcription; closed caption transcriptions; discriminative training; language models; large vocabulary continuous speech recognition; lightly supervised training; maximum likelihood training; maximum mutual information; minimum phone error; word error rate; Error analysis; Filtering; Maximum likelihood estimation; Mutual information; Parameter estimation; Radio broadcasting; Speech recognition; TV broadcasting; Training data; Vocabulary;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on
ISSN
1520-6149
Print_ISBN
0-7803-8484-9
Type
conf
DOI
10.1109/ICASSP.2004.1326091
Filename
1326091
Link To Document