DocumentCode
3431007
Title
An overhead reduction technique for mega-state compression schemes
Author
Bookstein, A. ; Klein, S.T. ; Raita, T.
Author_Institution
Chicago Univ., IL, USA
fYear
1997
fDate
25-27 Mar 1997
Firstpage
367
Lastpage
376
Abstract
Many of the most effective compression methods involve complicated models. Unfortunately, as model complexity increases, so does the cost of storing the model itself. This paper examines a method to reduce the amount of storage needed to represent a Markov model with an extended alphabet, by applying a clustering scheme that brings together similar states. Experiments run on a variety of large natural language texts show that much of the overhead of storing the model can be saved at the cost of a very small loss of compression efficiency
Keywords
Markov processes; data compression; natural languages; word processing; Markov model; clustering scheme; compression efficiency; experiments; extended alphabet; megastate compression schemes; model complexity; natural language texts; overhead reduction; storage reduction; Character generation; Context modeling; Costs; Data compression; Decoding; Frequency; History; Natural languages; State-space methods; Text recognition;
fLanguage
English
Publisher
ieee
Conference_Titel
Data Compression Conference, 1997. DCC '97. Proceedings
Conference_Location
Snowbird, UT
ISSN
1068-0314
Print_ISBN
0-8186-7761-9
Type
conf
DOI
10.1109/DCC.1997.582061
Filename
582061
Link To Document