Title :
Scaling laws for learning high-dimensional Markov forest distributions
Author :
Tan, Vincent Y F ; Anandkumar, Animashree ; Willsky, Alan S.
Author_Institution :
Stochastic Syst. Group, MIT, Cambridge, MA, USA
fDate :
Sept. 29 2010-Oct. 1 2010
Abstract :
The problem of learning forest-structured discrete graphical models from i.i.d. samples is considered. An algorithm based on pruning of the Chow-Liu tree through adaptive thresholding is proposed. It is shown that this algorithm is structurally consistent and the error probability of structure learning decays faster than any polynomial in the number of samples under fixed model size. For the high-dimensional scenario where the size of the model d and the number of edges k scale with the number of samples n, sufficient conditions on (n, d, k) are given for the algorithm to be structurally consistent. In addition, the extremal structures for learning are identified; we prove that the independent (resp. tree) model is the hardest (resp. easiest) to learn using the proposed algorithm in terms of error rates for structure learning.
Keywords :
Markov processes; error statistics; learning (artificial intelligence); polynomials; trees (mathematics); Chow-Liu tree; adaptive thresholding; error probability; forest structured discrete graphical models; high-dimensional Markov forest distributions; polynomial; pruning; scaling laws; structure learning; Error analysis; Error probability; Graphical models; Hidden Markov models; Joints; Markov processes; Mutual information; Forest distributions; Graphical models; Method of types; Structural consistency;
Conference_Titel :
Communication, Control, and Computing (Allerton), 2010 48th Annual Allerton Conference on
Conference_Location :
Allerton, IL
Print_ISBN :
978-1-4244-8215-3
DOI :
10.1109/ALLERTON.2010.5706977