DocumentCode :
1696031
Title :
Comparing RNNs and log-linear interpolation of improved skip-model on four Babel languages: Cantonese, Pashto, Tagalog, Turkish
Author :
Singh, Monika ; Klakow, Dietrich
Author_Institution :
Saarland Univ., Saarbrucken, Germany
fYear :
2013
Firstpage :
8416
Lastpage :
8420
Abstract :
Recurrent neural networks (RNNs) are a very recent technique to model long range dependencies in natural languages. They have clearly outperformed trigrams and other more advanced language modeling techniques by using non-linearly modeling long range dependencies. An alternative is to use log-linear interpolation of skip models (i.e. skip bigrams and skip trigrams). The method as such has been published earlier. In this paper we investigate the impact of different smoothing techniques on the skip models as a measure of their overall performance. One option is to use automatically trained distance clusters (both hard and soft) to increase robustness and to combat sparseness in the skip model. We also investigate alternative smoothing techniques on word level. For skip bigrams when skipping a small number of words Kneser-Ney smoothing (KN) is advantageous. For a larger number of words being skipped Dirichlet smoothing performs better. In order to exploit the advantages of both KN and Dirichlet smoothing we propose a new unified smoothing technique. Experiments are performed on four Babel languages: Cantonese, Pashto, Tagalog and Turkish. RNNs and log-linearly interpolated skip models are on par if the skip models are trained with standard smoothing techniques. Using the improved smoothing of the skip models along with distance clusters, we can clearly outperform RNNs by about 8-11 % in perplexity across all four languages.
Keywords :
interpolation; natural language processing; recurrent neural nets; smoothing methods; Babel language; Cantonese language; Dirichlet smoothing; KN smoothing technique; Kneser-Ney smoothing; Pashto language; RNN; Tagalog language; Turkish language; advanced language modeling technique; log-linear interpolation; recurrent neural network; skip-model; word level; Computational modeling; Context; Context modeling; Interpolation; Recurrent neural networks; Smoothing methods; Standards; RNNs; log-linear interpolation; skip models; smoothing; under researched languages;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
ISSN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2013.6639307
Filename :
6639307
Link To Document :
بازگشت