DocumentCode
639928
Title
Logarithmic Sobolev inequalities and strong data processing theorems for discrete channels
Author
Raginsky, Maxim
Author_Institution
Dept. of Electr. & Comput. Eng., Univ. of Illinois, Urbana, IL, USA
fYear
2013
fDate
7-12 July 2013
Firstpage
419
Lastpage
423
Abstract
The noisiness of a channel can be measured by comparing suitable functionals of the input and output distributions. For instance, if we fix a reference input distribution, then the worst-case ratio of output relative entropy to input relative entropy for any other input distribution is bounded by one, by the data processing theorem. However, for a fixed reference input distribution, this quantity may be strictly smaller than one, giving so-called strong data processing inequalities (SDPIs). This paper shows that the problem of determining both the best constant in an SDPI and any input distributions that achieve it can be addressed using so-called logarithmic Sobolev inequalities, which relate input relative entropy to certain measures of input-output correlation. Another contribution is a proof of equivalence between SDPIs and a limiting case of certain strong data processing inequalities for the Rényi divergence.
Keywords
entropy; Renyi divergence; SDPI; data processing theorem; discrete channels; fixed reference input distribution; logarithmic Sobolev inequalities; relative entropy; strong data processing inequalities; Correlation; Data processing; Entropy; Equations; Information theory; Limiting; Markov processes;
fLanguage
English
Publisher
ieee
Conference_Titel
Information Theory Proceedings (ISIT), 2013 IEEE International Symposium on
Conference_Location
Istanbul
ISSN
2157-8095
Type
conf
DOI
10.1109/ISIT.2013.6620260
Filename
6620260
Link To Document