Scaling laws for learning high-dimensional Markov forest distributions

The problem of learning forest-structured discrete graphical models from i.i.d. samples is considered. An algorithm based on pruning of the Chow-Liu tree through adaptive thresholding is proposed. It is shown that this algorithm is structurally consistent and the error probability of structure learn...

Full description

Bibliographic Details
Main Authors: Willsky, Alan S. (Contributor), Tan, Vincent Yan Fu (Contributor), Anandkumar, Animashree (Contributor)
Other Authors: Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science (Contributor), Massachusetts Institute of Technology. Laboratory for Information and Decision Systems (Contributor)
Format: Article
Language:English
Published: Institute of Electrical and Electronics Engineers (IEEE), 2012-10-04T13:43:41Z.
Subjects:
Online Access:Get fulltext
LEADER 02282 am a22002773u 4500
001 73590
042 |a dc 
100 1 0 |a Willsky, Alan S.  |e author 
100 1 0 |a Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science  |e contributor 
100 1 0 |a Massachusetts Institute of Technology. Laboratory for Information and Decision Systems  |e contributor 
100 1 0 |a Willsky, Alan S.  |e contributor 
100 1 0 |a Tan, Vincent Yan Fu  |e contributor 
100 1 0 |a Anandkumar, Animashree  |e contributor 
700 1 0 |a Tan, Vincent Yan Fu  |e author 
700 1 0 |a Anandkumar, Animashree  |e author 
245 0 0 |a Scaling laws for learning high-dimensional Markov forest distributions 
260 |b Institute of Electrical and Electronics Engineers (IEEE),   |c 2012-10-04T13:43:41Z. 
856 |z Get fulltext  |u http://hdl.handle.net/1721.1/73590 
520 |a The problem of learning forest-structured discrete graphical models from i.i.d. samples is considered. An algorithm based on pruning of the Chow-Liu tree through adaptive thresholding is proposed. It is shown that this algorithm is structurally consistent and the error probability of structure learning decays faster than any polynomial in the number of samples under fixed model size. For the high-dimensional scenario where the size of the model d and the number of edges k scale with the number of samples n, sufficient conditions on (n, d, k) are given for the algorithm to be structurally consistent. In addition, the extremal structures for learning are identified; we prove that the independent (resp. tree) model is the hardest (resp. easiest) to learn using the proposed algorithm in terms of error rates for structure learning. 
520 |a United States. Air Force Office of Scientific Research (Grant FA9559-08-1- 1080) 
520 |a United States. Army Research Office. Multidisciplinary University Research Initiative (Grant W911NF-06-1-0076) 
520 |a United States. Army Research Office. Multidisciplinary University Research Initiative (Grant FA9550-06-1-0324) 
520 |a Singapore. Agency for Science, Technology and Research 
546 |a en_US 
655 7 |a Article 
773 |t Proceedings of the 48th Annual Allerton Conference on Communication, Control, and Computing (Allerton), 2010