Title :
Scaling shrinkage-based language models
Author :
Chen, Stanley F. ; Mangu, Lidia ; Ramabhadran, Bhuvana ; Sarikaya, Ruhi ; Sethy, Abhinav
Author_Institution :
IBM T.J. Watson Res. Center, Yorktown Heights, NY, USA
fDate :
Nov. 13 2009-Dec. 17 2009
Abstract :
In we show that a novel class-based language model, Model M, and the method of regularized minimum discrimination information (rMDI) models outperform comparable methods on moderate amounts of Wall Street Journal data. Both of these methods are motivated by the observation that shrinking the sum of parameter magnitudes in an exponential language model tends to improve performance. In this paper, we investigate whether these shrinkage-based techniques also perform well on larger training sets and on other domains. First, we explain why good performance on large data sets is uncertain, by showing that gains relative to a baseline n-gram model tend to decrease as training set size increases. Next, we evaluate several methods for data/model combination with Model M and rMDI models on limited-scale domains, to uncover which techniques should work best on large domains. Finally, we apply these methods on a variety of medium-to-large-scale domains covering several languages, and show that Model M consistently provides significant gains over existing language models for state-of-the-art systems in both speech recognition and machine translation.
Keywords :
language translation; natural language processing; speech recognition; language model; machine translation; regularized minimum discrimination information models; shrinkage-based techniques; speech recognition; Acoustic testing; Automatic speech recognition; Interpolation; Large-scale systems; Natural languages; Performance gain; Predictive models; Speech recognition; Training data;
Conference_Titel :
Automatic Speech Recognition & Understanding, 2009. ASRU 2009. IEEE Workshop on
Conference_Location :
Merano
Print_ISBN :
978-1-4244-5478-5
Electronic_ISBN :
978-1-4244-5479-2
DOI :
10.1109/ASRU.2009.5373380