DocumentCode
1779527
Title
Exact common information
Author
Kumar, G. Ramesh ; Cheuk Ting Li ; El Gamal, Abbas
Author_Institution
Electr. Eng., Stanford Univ., Stanford, CA, USA
fYear
2014
fDate
June 29 2014-July 4 2014
Firstpage
161
Lastpage
165
Abstract
This paper introduces the notion of exact common information, which is the minimum description length of the common randomness needed for the exact distributed generation of two correlated random variables (X, Y). We introduce the quantity G(X; Y) = minX→W→Y H(W) as a natural bound on the exact common information and study its properties and computation. We then introduce the exact common information rate, which is the minimum description rate of the common randomness for the exact generation of a 2-DMS (X, Y). We give a multiletter characterization for it as the limit G̅(X; Y) = limn→∞(1/n)G(Xn; Yn). While in general G̅(X; Y) is greater than or equal to the Wyner common information, we show that they are equal for the Symmetric Binary Erasure Source. We do not know, however, if the exact common information rate has a single letter characterization in general.
Keywords
approximation theory; correlation theory; entropy; Wyner common information; correlated random variables; exact common information rate; exact distributed generation; exact generation; minimum description length; minimum description rate; multiletter characterization; single letter characterization; symmetric binary erasure source; Decoding; Distributed power generation; Educational institutions; Entropy; Information rates; Random variables;
fLanguage
English
Publisher
ieee
Conference_Titel
Information Theory (ISIT), 2014 IEEE International Symposium on
Conference_Location
Honolulu, HI
Type
conf
DOI
10.1109/ISIT.2014.6874815
Filename
6874815
Link To Document