Title :
An improved recurrent neural network language model with context vector features
Author :
Jian Zhang ; Dan Qu ; Zhen Li
Author_Institution :
Nat. Digital Switching Syst. Eng. & Technol. R&D, Center Zhengzhou, Zhengzhou, China
Abstract :
Recurrent neural network language models have solved the problems of data sparseness and dimensionality disaster which exist in traditional N-gram models. RNNLMs have recently demonstrated state-of-the-art performance in speech recognition, machine translation and other tasks. In this paper, we improve the model performance by providing contextual word vectors in association with RNNLMs. This method can reinforce the ability of learning long-distance information using vectors training from Skip-gram model. The experimental results show that the proposed method can improve the perplexity performance significantly on Penn Treebank data. And we further apply the models to speech recognition task on the Wall Street Journal corpora, where we achieve obvious improvements in word-error-rate.
Keywords :
recurrent neural nets; speech recognition; N-gram models; Penn Treebank data; RNNLM; Wall Street Journal corpora; context vector features; contextual word vectors; data dimensionality; data sparseness; long-distance information learning; machine translation; perplexity performance; recurrent neural network language model; speech recognition task; Computational modeling; Context; Context modeling; Hidden Markov models; Neurons; Recurrent neural networks; Vectors; Language Model; Recurrent Neural Network; Skip-gram; Speech Recognition;
Conference_Titel :
Software Engineering and Service Science (ICSESS), 2014 5th IEEE International Conference on
Conference_Location :
Beijing
Print_ISBN :
978-1-4799-3278-8
DOI :
10.1109/ICSESS.2014.6933694