Large Margin Hierarchical Classification

Abstract

We present an algorithmic framework for supervised classification learningwhere the set of labels is organized in a predefined hierarchical structure. This structure is encoded by a rooted tree which induces a metric over thelabel set. Our approach combines ideas from large margin kernel methods and Bayesian analysis. Following the large margin principle, we associate aprototype with each label in the tree and formulate the learning task as anoptimization problem with varying margin constraints. In the spirit of Bayesian methods, we impose similarity requirements between the prototypescorresponding to adjacent labels in the hierarchy. We describe new online andbatch algorithms for solving the constrained optimization problem. We derive aworst case loss-bound for the online algorithm and provide generalizationanalysis for its batch counterpart. We demonstrate the merits of our approachwith a series of experiments on synthetic, text and speech data.

Cite

Text

Dekel et al. "Large Margin Hierarchical Classification." International Conference on Machine Learning, 2004. doi:10.1145/1015330.1015374

Markdown

[Dekel et al. "Large Margin Hierarchical Classification." International Conference on Machine Learning, 2004.](https://mlanthology.org/icml/2004/dekel2004icml-large/) doi:10.1145/1015330.1015374

BibTeX

@inproceedings{dekel2004icml-large,
  title     = {{Large Margin Hierarchical Classification}},
  author    = {Dekel, Ofer and Keshet, Joseph and Singer, Yoram},
  booktitle = {International Conference on Machine Learning},
  year      = {2004},
  doi       = {10.1145/1015330.1015374},
  url       = {https://mlanthology.org/icml/2004/dekel2004icml-large/}
}