DocumentCode
1106784
Title
Estimating Mutual Information Via Kolmogorov Distance
Author
Zhang, Zhengmin
Author_Institution
Carleton Univ., Ottawa
Volume
53
Issue
9
fYear
2007
Firstpage
3280
Lastpage
3282
Abstract
By use of a coupling technique, two inequalities are established which set upper bounds to the mutual information of finite discrete random variables in terms of the Kolmogorov distance (variational distance).
Keywords
entropy; random processes; Kolmogorov distance; Shannon entropy; coupling technique; finite discrete random variables; set upper bounds; variational distance; Entropy; Information theory; Mathematics; Mutual coupling; Mutual information; Probability distribution; Random variables; Statistical distributions; Testing; Upper bound; Kolmogorov distance; Shannon entropy; mutual information;
fLanguage
English
Journal_Title
Information Theory, IEEE Transactions on
Publisher
ieee
ISSN
0018-9448
Type
jour
DOI
10.1109/TIT.2007.903122
Filename
4294175
Link To Document