Simultaneous learning of trees and representations for extreme classification and density estimation

We consider multi-class classification where the predictor has a hierarchical structure that allows for a very large number of labels both at train and test time. The predictive power of such models can heavily depend on the structure of the tree, and although past work showed how to learn the tree...

Full description

Bibliographic Details
Main Author: Sontag, David Alexander (Author)
Other Authors: Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science (Contributor)
Format: Article
Language:English
Published: International Machine Learning Society, 2021-04-27T17:45:25Z.
Subjects:
Online Access:Get fulltext
LEADER 01661 am a22001573u 4500
001 130531
042 |a dc 
100 1 0 |a Sontag, David Alexander  |e author 
100 1 0 |a Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science  |e contributor 
245 0 0 |a Simultaneous learning of trees and representations for extreme classification and density estimation 
260 |b International Machine Learning Society,   |c 2021-04-27T17:45:25Z. 
856 |z Get fulltext  |u https://hdl.handle.net/1721.1/130531 
520 |a We consider multi-class classification where the predictor has a hierarchical structure that allows for a very large number of labels both at train and test time. The predictive power of such models can heavily depend on the structure of the tree, and although past work showed how to learn the tree structure, it expected that the feature vectors remained static. We provide a novel algorithm to simultaneously perform representation learning for the input data and learning of the hierarchical predictor. Our approach optimizes an objective function which favors balanced and easily-separable multi-way node partitions. We theoretically analyze this objective, showing that it gives rise to a boosting style property and a bound on classification error. We next show how to extend the algorithm to conditional density estimation. We empirically validate both variants of the algorithm on text classification and language modeling, respectively, and show that they compare favorably to common baselines in terms of accuracy and running time. 
546 |a en 
655 7 |a Article 
773 |t Proceedings of Machine Learning Research