DocumentCode :
3162043
Title :
Distributed acoustic modeling with back-off n-grams
Author :
Chelba, Ciprian ; Xu, Peng ; Pereira, Fernando ; Richardson, Thomas
Author_Institution :
Google, Inc., Mountain View, CA, USA
fYear :
2012
fDate :
25-30 March 2012
Firstpage :
4129
Lastpage :
4132
Abstract :
The paper proposes an approach to acoustic modeling that borrows from n-gram language modeling in an attempt to scale up both the amount of training data and model size (as measured by the number of parameters in the model) to approximately 100 times larger than current sizes used in ASR. Dealing with unseen phonetic contexts is accomplished using the familiar back-off technique used in language modeling due to implementation simplicity. The new acoustic model is estimated and stored using the Map-Reduce distributed computing infrastructure. Speech recognition experiments are carried out in an N-best rescoring framework for Google Voice Search. 87,000 hours of training data is obtained in an unsupervised fashion by filtering utterances in Voice Search logs on ASR confidence. The resulting models are trained using maximum likelihood and contain 20-40 million Gaussians. They achieve relative reductions in WER of 11% and 6% over first-pass models trained using maximum likelihood, and boosted MMI, respectively.
Keywords :
maximum likelihood estimation; speech processing; speech recognition; ASR; Google voice search; MapReduce distributed computing infrastructure; WER; back-off n-gram language model; boosted MMI; distributed acoustic modeling; maximum likelihood method; model size; training data; unseen phonetic contexts; Acoustics; Context; Data models; Hidden Markov models; Speech; Training; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
Conference_Location :
Kyoto
ISSN :
1520-6149
Print_ISBN :
978-1-4673-0045-2
Electronic_ISBN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2012.6288827
Filename :
6288827
Link To Document :
بازگشت