Title :
EM mixture model probability table compression
Author :
Cho, Sung-Jung ; Perrone, Michael ; Ratzlaff, Eugene
Author_Institution :
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
Abstract :
This paper presents a new probability table compression method based on mixture models applied to N-tuple recognizers. Joint probability tables are modeled by lower dimensional probability mixtures and their mixture coefficients. The maximum likelihood parameters of the mixture models are trained by the expectation-maximization (EM) algorithm and quantized to one byte integers. The probability elements which mixture models do not estimate reliably are kept separately. Experimental results with on-line handwritten UNIPEN digits show that the total memory size of an N-tuple recognizer is reduced from 11.8 Mbytes to 0.55 Mbytes, while the recognition rate drops from 97.7% to 97.5%.
Keywords :
data compression; handwriting recognition; maximum likelihood estimation; optimisation; probability; 0.55 Mbyte; EM algorithm; EM mixture model probability table compression; N-tuple recognizers; expectation-maximization algorithm; handwriting recognition systems; joint probability tables; maximum likelihood parameters; memory size; memory usage; mixture coefficients; mixture models; probability mixtures; recognition rate; Character generation; Character recognition; Handheld computers; Handwriting recognition; Image coding; Maximum likelihood estimation; Personal digital assistants; Quantization; Real time systems; Training data;
Conference_Titel :
Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03). 2003 IEEE International Conference on
Print_ISBN :
0-7803-7663-3
DOI :
10.1109/ICASSP.2003.1202491