Shrinking the Size of Deep Extreme Multi-Label Classification
Abstract
Training deep classifiers for Extreme Multi-Label Classification (XMC) is difficult due to the computational and memory costs caused by extremely large label sets. Traditionally, the final output layer of these deep classifiers scales linearly with the size of the label set, which is often in the millions for realistic settings. Reducing the size of deep classifiers for XMC is necessary to (i) train them more efficiently and (ii) deploy them within memory-constrained systems, such as mobile devices. We address the current limitations of deep classifiers by proposing a novel XMC method, DECLARE: Deep Extreme Compressed Labeling And Recovery Estimation. DECLARE compresses the labels into a smaller dimension, which reduces the training time and model-storage size. DECLARE retains enough information to recover the most-likely predicted labels in the original label space. Empirically, DECLARE compresses labels by up to 99.975% while outperforming uncompressed performance.
Cite
Text
Bornstein et al. "Shrinking the Size of Deep Extreme Multi-Label Classification." NeurIPS 2024 Workshops: Compression, 2024.Markdown
[Bornstein et al. "Shrinking the Size of Deep Extreme Multi-Label Classification." NeurIPS 2024 Workshops: Compression, 2024.](https://mlanthology.org/neuripsw/2024/bornstein2024neuripsw-shrinking/)BibTeX
@inproceedings{bornstein2024neuripsw-shrinking,
title = {{Shrinking the Size of Deep Extreme Multi-Label Classification}},
author = {Bornstein, Marco and Rabbani, Tahseen and Gravelle, Brian Joseph and Huang, Furong},
booktitle = {NeurIPS 2024 Workshops: Compression},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/bornstein2024neuripsw-shrinking/}
}