Author :
Sagae, K. ; Lehr, M. ; Prud´hommeaux, E. ; Xu, P. ; Glenn, N. ; Karakos, D. ; Khudanpur, S. ; Roark, B. ; Saraçlar, M. ; Shafran, I. ; Bikel, D. ; Callison-Burch, C. ; Cao, Y. ; Hall, K. ; Hasler, E. ; Koehn, P. ; Lopez, A. ; Post, M. ; Riley, D.
Abstract :
This paper investigates semi-supervised methods for discriminative language modeling, whereby n-best lists are “hallucinated” for given reference text and are then used for training n-gram language models using the perceptron algorithm. We perform controlled experiments on a very strong baseline English CTS system, comparing three methods for simulating ASR output, and compare the results with training with “real” n-best list output from the baseline recognizer. We find that methods based on extracting phrasal cohorts - similar to methods from machine translation for extracting phrase tables - yielded the largest gains of our three methods, achieving over half of the WER reduction of the fully supervised methods.
Keywords :
language translation; natural language processing; English CTS system; baseline recognizer; discriminative language modeling; hallucinated n-best lists; machine translation; perceptron algorithm; reference text; semisupervised method; simulating ASR output; training n-gram language model; Data models; Hidden Markov models; Speech; Speech recognition; Training; Training data; Transducers; automatic speech recognition; discriminative training; language modeling; semi-supervised methods;