DocumentCode
2964220
Title
Dynamic network decoding revisited
Author
Soltau, Hagen ; Saon, George
Author_Institution
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
fYear
2009
fDate
Nov. 13 2009-Dec. 17 2009
Firstpage
276
Lastpage
281
Abstract
We present a dynamic network decoder capable of using large cross-word context models and large n-gram histories. Our method for constructing the search network is designed to process large cross-word context models very efficiently and we address the optimization of the search network to minimize any overhead during run-time for the dynamic network decoder. The search procedure uses the full LM history for lookahead, and path recombination is done as early as possible. In our systematic comparison to a static FSM based decoder, we find the dynamic decoder can run at comparable speed as the static decoder when large language models are used, while the static decoder performs best for small language models. We discuss the use of very large vocabularies of up to 2.5 million words for both decoding approaches and analyze the effect of weak acoustic models for pruning.
Keywords
decoding; finite state machines; acoustic models; cross-word context models; dynamic network decoding; finite state machines; n-gram language model; static FSM; Computer networks; Context modeling; Costs; Decoding; Design optimization; Hidden Markov models; History; Process design; Runtime; Vocabulary;
fLanguage
English
Publisher
ieee
Conference_Titel
Automatic Speech Recognition & Understanding, 2009. ASRU 2009. IEEE Workshop on
Conference_Location
Merano
Print_ISBN
978-1-4244-5478-5
Electronic_ISBN
978-1-4244-5479-2
Type
conf
DOI
10.1109/ASRU.2009.5372904
Filename
5372904
Link To Document