More data means less inference: A pseudo-max approach to structured learning

The problem of learning to predict structured labels is of key importance in many applications. However, for general graph structure both learning and inference in this setting are intractable. Here we show that it is possible to circumvent this difficulty when the input distribution is rich enough...

Full description

Bibliographic Details
Main Authors: Sontag, David (Author), Meshi, Ofer (Author), Jaakkola, Tommi S. (Contributor), Globerson, Amir (Author)
Other Authors: Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory (Contributor), Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science (Contributor)
Format: Article
Language:English
Published: Neural Information Processing Systems Foundation, 2011-07-06T14:56:22Z.
Subjects:
Online Access:Get fulltext
LEADER 01707 am a22002653u 4500
001 64743
042 |a dc 
100 1 0 |a Sontag, David  |e author 
100 1 0 |a Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory  |e contributor 
100 1 0 |a Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science  |e contributor 
100 1 0 |a Jaakkola, Tommi S.  |e contributor 
100 1 0 |a Jaakkola, Tommi S.  |e contributor 
700 1 0 |a Meshi, Ofer  |e author 
700 1 0 |a Jaakkola, Tommi S.  |e author 
700 1 0 |a Globerson, Amir  |e author 
245 0 0 |a More data means less inference: A pseudo-max approach to structured learning 
260 |b Neural Information Processing Systems Foundation,   |c 2011-07-06T14:56:22Z. 
856 |z Get fulltext  |u http://hdl.handle.net/1721.1/64743 
520 |a The problem of learning to predict structured labels is of key importance in many applications. However, for general graph structure both learning and inference in this setting are intractable. Here we show that it is possible to circumvent this difficulty when the input distribution is rich enough via a method similar in spirit to pseudo-likelihood. We show how our new method achieves consistency, and illustrate empirically that it indeed performs as well as exact methods when sufficiently large training sets are used. 
520 |a United States-Israel Binational Science Foundation (Grant 2008303) 
520 |a Google (Firm) (Research Grant) 
520 |a Google (Firm) (PhD Fellowship) 
546 |a en_US 
655 7 |a Article 
773 |t Poster Session paper of the Twenty-Fourth Annual Conference on Neural Information Processing Systems, NIPS 2010