DocumentCode :
38210
Title :
How Suboptimal Is the Shannon Code?
Author :
Narimani, Hamed ; Khosravifard, Mohammadali ; Gulliver, T. Aaron
Author_Institution :
Dept. of Electr. & Comput. Eng., Isfahan Univ. of Technol., Isfahan, Iran
Volume :
59
Issue :
1
fYear :
2013
fDate :
Jan. 2013
Firstpage :
458
Lastpage :
471
Abstract :
In order to determine how suboptimal the Shannon code is, one should compare its performance with that of the optimal code, i.e., the corresponding Huffman code, in some sense. It is well known that in the worst case the redundancy of both the Shannon and Huffman codes can be arbitrarily close to 1. Beyond this worst case viewpoint, very little is known. In this paper, we compare the performance of these codes from an average point of view. The redundancy is considered as a random variable on the set of all sources with n symbols and its average is evaluated. It is shown that the average redundancy of the Shannon code is very close to 0.5 bits, whereas the average redundancy of the Huffman code is less than n-1(1+ln n)+0.086 bits . It is also proven that the variance of the redundancy of the Shannon code tends to zero as n increases. Therefore, for sources with alphabet size n, the redundancy of the Shannon code is approximately 0.5 bits with probability approaching 1 as n→ ∞.
Keywords :
Huffman codes; information theory; Huffman code; Shannon code; optimal code; Channel coding; Entropy; Probability density function; Random variables; Redundancy; Upper bound; Average redundancy; Huffman code; Shannon code;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2012.2216975
Filename :
6293889
Link To Document :
بازگشت