DocumentCode :
1203452
Title :
The empirical distribution of rate-constrained source codes
Author :
Weissman, Tsachy ; Ordentlich, Erik
Author_Institution :
Electr. Eng. Dept., Stanford Univ., CA, USA
Volume :
51
Issue :
11
fYear :
2005
Firstpage :
3718
Lastpage :
3733
Abstract :
Let X = (X1,...) be a stationary ergodic finite-alphabet source, Xn denote its first n symbols, and Yn be the codeword assigned to Xn by a lossy source code. The empirical kth-order joint distribution Qˆk[Xn,Yn(xk,yk) is defined as the frequency of appearances of pairs of k-strings (xk,yk) along the pair (Xn,Yn). Our main interest is in the sample behavior of this (random) distribution. Letting I(Qk) denote the mutual information I(Xk;Yk) when (Xk,Yk)∼Qk we show that for any (sequence of) lossy source code(s) of rate ≤R lim supn→∞(1/k)I(Qˆk[Xn,Yn) ≤R+(1/k)H (X1k)-H~(X) a.s. where H~(X) denotes the entropy rate of X. This is shown to imply, for a large class of sources including all independent and identically distributed (i.i.d.). sources and all sources satisfying the Shannon lower bound with equality, that for any sequence of codes which is good in the sense of asymptotically attaining a point on the rate distortion curve Qˆk[Xn,YndP(Xk,Y~k) a.s. whenever P(Xk,Yk) is the unique distribution attaining the minimum in the definition of the kth-order rate distortion function. Consequences of these results include a new proof of Kieffer\´s sample converse to lossy source coding, as well as performance bounds for compression-based denoisers.
Keywords :
combined source-channel coding; constraint theory; data compression; entropy codes; rate distortion theory; sequential codes; Kieffer´s compression-based denoiser; Shannon lower bound; code sequence; codeword assignment; empirical distribution; empirical kth-order joint distribution; entropy rate; i.i.d. source; independent identically distributed source; mutual information; rate distortion curve; rate-constrained source code; sample conversion; stationary ergodic finite-alphabet; Entropy; Frequency; Heart; Laboratories; Mutual information; Noise reduction; Performance loss; Rate distortion theory; Rate-distortion; Source coding; Compression-based denoising; Shannon lower bound; empirical distribution; good codes; lossy source codes; pointwise bounds; strong converse;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2005.856982
Filename :
1522635
Link To Document :
بازگشت