Title :
Structured Output Layer neural network language model
Author :
Le, Hai-Son ; Oparin, Ilya ; Allauzen, Alexandre ; Gauvain, Jean-Luc ; Yvon, François
Abstract :
This paper introduces a new neural network language model (NNLM) based on word clustering to structure the output vocabulary: Structured Output Layer NNLM. This model is able to handle vocabularies of arbitrary size, hence dispensing with the design of short-lists that are commonly used in NNLMs. Several softmax layers replace the standard output layer in this model. The output structure depends on the word clustering which uses the continuous word representation induced by a NNLM. The GALE Mandarin data was used to carry out the speech-to-text experiments and evaluate the NNLMs. On this data the well tuned baseline system has a character error rate under 10%. Our model achieves consistent improvements over the combination of an n-gram model and classical short-list NNLMs both in terms of perplexity and recognition accuracy.
Keywords :
natural language processing; neural nets; speech recognition; speech synthesis; text analysis; GALE Mandarin data; character error rate; n-gram model; softmax layers; speech recognition accuracy; speech-to-text experiments; standard output layer; structured output layer NNLM; structured output layer neural network language model; word clustering; Adaptation models; Artificial neural networks; Computational modeling; Context; Speech recognition; Training; Vocabulary; Automatic Speech Recognition; Neural Network Language Model; Speech-To-Text;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on
Conference_Location :
Prague
Print_ISBN :
978-1-4577-0538-0
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2011.5947610