Title :
Continuous space language modeling techniques
Author :
Sarikaya, Ruhi ; Emami, Ahmad ; Afify, Mohamed ; Ramabhadran, Bhuvana
Author_Institution :
IBM T.J. Watson Res. Center, Yorktown Heights, NY, USA
Abstract :
This paper focuses on comparison of two continuous space language modeling techniques, namely Tied-Mixture Language modeling (TMLM) and Neural Network Based Language Modeling (NNLM). Additionally, we report on using alternative feature representations for words and histories used in TMLM. Besides bigram co-occurrence based features we consider using NNLM based input features for training TMLMs. We also describe how we improve certain steps in building TMLMs. We demonstrate that TMLMs provide significant improvements of over 16% relative and 10% relative in Character Error Rate (CER) for Mandarin speech recognition, over the trigram and NNLM models, respectively in a speech to speech translation task.
Keywords :
natural language processing; neural nets; speech processing; speech recognition; Mandarin speech recognition; bigram co-occurrence; character error rate; continuous space language modeling techniques; feature representations; neural network based language modeling; speech-to-speech translation task; tied mixture language modeling; Degradation; Error analysis; Hidden Markov models; History; Maximum likelihood decoding; Natural language processing; Natural languages; Neural networks; Speech recognition; Training data; Continuous Space Modeling; Language Modeling; NNLM; Tied-Mixture Modeling;
Conference_Titel :
Acoustics Speech and Signal Processing (ICASSP), 2010 IEEE International Conference on
Conference_Location :
Dallas, TX
Print_ISBN :
978-1-4244-4295-9
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2010.5495009