Title :
Empirical study of neural network language models for Arabic speech recognition
Author :
Emami, Ahmad ; Mangu, Lidia
Author_Institution :
IBM T J Watson Res. Center, Yorktown Heights
Abstract :
In this paper we investigate the use of neural network language models for Arabic speech recognition. By using a distributed representation of words, the neural network model allows for more robust generalization and is better able to fight the data sparseness problem. We investigate different configurations of the neural probabilistic model, experimenting with such parameters as N-gram order, output vocabulary, normalization method, and model size and parameters. Experiments were carried out on Arabic broadcast news and broadcast conversations data and the optimized neural network language models showed significant improvements over the baseline N-gram model.
Keywords :
natural language processing; neural nets; probability; speech recognition; Arabic broadcast news; Arabic speech recognition; data sparseness problem; neural network language models; neural probabilistic model; normalization method; robust generalization; Broadcasting; History; Natural languages; Neural networks; Polynomials; Probability; Robustness; Smoothing methods; Speech recognition; Vocabulary; Language Modeling; Neural Networks; Speech Recognition;
Conference_Titel :
Automatic Speech Recognition & Understanding, 2007. ASRU. IEEE Workshop on
Conference_Location :
Kyoto
Print_ISBN :
978-1-4244-1746-9
Electronic_ISBN :
978-1-4244-1746-9
DOI :
10.1109/ASRU.2007.4430100