• DocumentCode
    395318
  • Title

    EM mixture model probability table compression

  • Author

    Cho, Sung-Jung ; Perrone, Michael ; Ratzlaff, Eugene

  • Author_Institution
    IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
  • Volume
    2
  • fYear
    2003
  • fDate
    6-10 April 2003
  • Abstract
    This paper presents a new probability table compression method based on mixture models applied to N-tuple recognizers. Joint probability tables are modeled by lower dimensional probability mixtures and their mixture coefficients. The maximum likelihood parameters of the mixture models are trained by the expectation-maximization (EM) algorithm and quantized to one byte integers. The probability elements which mixture models do not estimate reliably are kept separately. Experimental results with on-line handwritten UNIPEN digits show that the total memory size of an N-tuple recognizer is reduced from 11.8 Mbytes to 0.55 Mbytes, while the recognition rate drops from 97.7% to 97.5%.
  • Keywords
    data compression; handwriting recognition; maximum likelihood estimation; optimisation; probability; 0.55 Mbyte; EM algorithm; EM mixture model probability table compression; N-tuple recognizers; expectation-maximization algorithm; handwriting recognition systems; joint probability tables; maximum likelihood parameters; memory size; memory usage; mixture coefficients; mixture models; probability mixtures; recognition rate; Character generation; Character recognition; Handheld computers; Handwriting recognition; Image coding; Maximum likelihood estimation; Personal digital assistants; Quantization; Real time systems; Training data;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03). 2003 IEEE International Conference on
  • ISSN
    1520-6149
  • Print_ISBN
    0-7803-7663-3
  • Type

    conf

  • DOI
    10.1109/ICASSP.2003.1202491
  • Filename
    1202491