Learning Taxonomy Adaptation in Large-Scale Classification
Abstract
In this paper, we study flat and hierarchical classification strategies in the context of large-scale taxonomies. Addressing the problem from a learning-theoretic point of view, we first propose a multi-class, hierarchical data dependent bound on the generalization error of classifiers deployed in large-scale taxonomies. This bound provides an explanation to several empirical results reported in the literature, related to the performance of flat and hierarchical classifiers. Based on this bound, we also propose a technique for modifying a given taxonomy through pruning, that leads to a lower value of the upper bound as compared to the original taxonomy. We then present another method for hierarchy pruning by studying approximation error of a family of classifiers, and derive from it features used in a meta-classifier to decide which nodes to prune. We finally illustrate the theoretical developments through several experiments conducted on two widely used taxonomies.
Cite
Text
Babbar et al. "Learning Taxonomy Adaptation in Large-Scale Classification." Journal of Machine Learning Research, 2016.Markdown
[Babbar et al. "Learning Taxonomy Adaptation in Large-Scale Classification." Journal of Machine Learning Research, 2016.](https://mlanthology.org/jmlr/2016/babbar2016jmlr-learning/)BibTeX
@article{babbar2016jmlr-learning,
title = {{Learning Taxonomy Adaptation in Large-Scale Classification}},
author = {Babbar, Rohit and Partalas, Ioannis and Gaussier, Eric and Amini, Massih-Reza and Amblard, Cécile},
journal = {Journal of Machine Learning Research},
year = {2016},
pages = {1-37},
volume = {17},
url = {https://mlanthology.org/jmlr/2016/babbar2016jmlr-learning/}
}