DocumentCode :
28704
Title :
Universal Estimation of Directed Information
Author :
Jiantao Jiao ; Permuter, Haim H. ; Lei Zhao ; Young-Han Kim ; Weissman, Tsachy
Author_Institution :
Dept. of Electr. Eng., Stanford Univ., Stanford, CA, USA
Volume :
59
Issue :
10
fYear :
2013
fDate :
Oct. 2013
Firstpage :
6220
Lastpage :
6242
Abstract :
Four estimators of the directed information rate between a pair of jointly stationary ergodic finite-alphabet processes are proposed, based on universal probability assignments. The first one is a Shannon-McMillan-Breiman-type estimator, similar to those used by Verdú in 2005 and Cai in 2006 for estimation of other information measures. We show the almost sure and L1 convergence properties of the estimator for any underlying universal probability assignment. The other three estimators map universal probability assignments to different functionals, each exhibiting relative merits such as smoothness, nonnegativity, and boundedness. We establish the consistency of these estimators in almost sure and L1 senses, and derive near-optimal rates of convergence in the minimax sense under mild conditions. These estimators carry over directly to estimating other information measures of stationary ergodic finite-alphabet processes, such as entropy rate and mutual information rate, with near-optimal performance and provide alternatives to classical approaches in the existing literature. Guided by these theoretical results, the proposed estimators are implemented using the context-tree weighting algorithm as the universal probability assignment. Experiments on synthetic and real data are presented, demonstrating the potential of the proposed schemes in practice and the utility of directed information estimation in detecting and measuring causal influence and delay.
Keywords :
causality; delays; entropy; estimation theory; probability; Shannon-McMillan-Breiman-type estimator; causal influence measurement; context-tree weighting algorithm; delay; directed information rate estimation; entropy rate; mutual information rate; near-optimal performance; stationary ergodic finite-alphabet process; universal probability assignment; Context; Convergence; Educational institutions; Entropy; Estimation; Information rates; Q measurement; Causal influence; context-tree weighting (CTW); directed information; rate of convergence; universal probability assignment;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2013.2267934
Filename :
6555871
Link To Document :
بازگشت