• DocumentCode
    3431007
  • Title

    An overhead reduction technique for mega-state compression schemes

  • Author

    Bookstein, A. ; Klein, S.T. ; Raita, T.

  • Author_Institution
    Chicago Univ., IL, USA
  • fYear
    1997
  • fDate
    25-27 Mar 1997
  • Firstpage
    367
  • Lastpage
    376
  • Abstract
    Many of the most effective compression methods involve complicated models. Unfortunately, as model complexity increases, so does the cost of storing the model itself. This paper examines a method to reduce the amount of storage needed to represent a Markov model with an extended alphabet, by applying a clustering scheme that brings together similar states. Experiments run on a variety of large natural language texts show that much of the overhead of storing the model can be saved at the cost of a very small loss of compression efficiency
  • Keywords
    Markov processes; data compression; natural languages; word processing; Markov model; clustering scheme; compression efficiency; experiments; extended alphabet; megastate compression schemes; model complexity; natural language texts; overhead reduction; storage reduction; Character generation; Context modeling; Costs; Data compression; Decoding; Frequency; History; Natural languages; State-space methods; Text recognition;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Data Compression Conference, 1997. DCC '97. Proceedings
  • Conference_Location
    Snowbird, UT
  • ISSN
    1068-0314
  • Print_ISBN
    0-8186-7761-9
  • Type

    conf

  • DOI
    10.1109/DCC.1997.582061
  • Filename
    582061