• DocumentCode
    3642153
  • Title

    Extensions of recurrent neural network language model

  • Author

    Tomáš Mikolov;Stefan Kombrink;Lukáš Burget;Jan Černocký;Sanjeev Khudanpur

  • Author_Institution
    Brno University of Technology, Speech@FIT, Czech Republic
  • fYear
    2011
  • fDate
    5/1/2011 12:00:00 AM
  • Firstpage
    5528
  • Lastpage
    5531
  • Abstract
    We present several modifications of the original recurrent neural net work language model (RNN LM). While this model has been shown to significantly outperform many competitive language modeling techniques in terms of accuracy, the remaining problem is the computational complexity. In this work, we show approaches that lead to more than 15 times speedup for both training and testing phases. Next, we show importance of using a backpropagation through time algorithm. An empirical comparison with feedforward networks is also provided. In the end, we discuss possibilities how to reduce the amount of parameters in the model. The resulting RNN model can thus be smaller, faster both during training and testing, and more accurate than the basic one.
  • Keywords
    "Recurrent neural networks","Artificial neural networks","Training","Computational modeling","Vocabulary","Backpropagation","Probability distribution"
  • Publisher
    ieee
  • Conference_Titel
    Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on
  • ISSN
    1520-6149
  • Print_ISBN
    978-1-4577-0538-0
  • Electronic_ISBN
    2379-190X
  • Type

    conf

  • DOI
    10.1109/ICASSP.2011.5947611
  • Filename
    5947611