DocumentCode
3265293
Title
The entropy of English using PPM-based models
Author
Teahan, W.J. ; Cleary, John G.
Author_Institution
Dept. of Comput. Sci., Waikato Univ., Hamilton, New Zealand
fYear
1996
fDate
Mar/Apr 1996
Firstpage
53
Lastpage
62
Abstract
The purpose of this paper is to show that the difference between the best machine models and human models is smaller than might be indicated by the previous results. This follows from a number of observations: firstly, the original human experiments used only 27 character English (letters plus space) against full 128 character ASCII text for most computer experiments; secondly, using large amounts of priming text substantially improves the PPM´s performance; and thirdly, the PPM algorithm can be modified to perform better for English text. The result of this is a machine performance down to 1.46 bit per character. The problem of estimating the entropy of English is discussed. The importance of training text for PPM is demonstrated, showing that its performance can be improved by “adjusting” the alphabet used. The results based on these improvements are then given, with compression down to 1.46 bpc
Keywords
data compression; entropy; speech processing; ASCII text; English; English text; PPM algorithm; PPM based models; alphabet; compression; computer experiments; entropy; human experiments; human models; letters; machine models; machine performance; priming text; training text; Computer science; Context modeling; Cryptography; Entropy; Humans; Natural languages; Optical character recognition software; Speech recognition; Statistics; Upper bound;
fLanguage
English
Publisher
ieee
Conference_Titel
Data Compression Conference, 1996. DCC '96. Proceedings
Conference_Location
Snowbird, UT
ISSN
1068-0314
Print_ISBN
0-8186-7358-3
Type
conf
DOI
10.1109/DCC.1996.488310
Filename
488310
Link To Document