ST$_k$: A Scalable Module for Solving Top-K Problems
Abstract
The cost of ranking becomes significant in the new stage of deep learning. We propose ST$_k$, a fully differentiable module with a single trainable parameter, designed to solve the Top-k problem without requiring additional time or GPU memory. Due to its fully differentiable nature, ST$_k$ can be embedded end-to-end into neural networks and optimize the Top-k problems within a unified computational graph. We apply ST$_k$ to the Average Top-k Loss (AT$_k$), which inherently faces a Top-k problem. The proposed ST$_k$ Loss outperforms AT$_k$ Loss and achieves the best average performance on multiple benchmarks, with the lowest standard deviation. With the assistance of ST$_k$ Loss, we surpass the state-of-the-art (SOTA) on both CIFAR-100-LT and Places-LT leaderboards.
Cite
Text
Xia et al. "ST$_k$: A Scalable Module for Solving Top-K Problems." Neural Information Processing Systems, 2024. doi:10.52202/079017-0534Markdown
[Xia et al. "ST$_k$: A Scalable Module for Solving Top-K Problems." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/xia2024neurips-st/) doi:10.52202/079017-0534BibTeX
@inproceedings{xia2024neurips-st,
title = {{ST$_k$: A Scalable Module for Solving Top-K Problems}},
author = {Xia, Hanchen and Liu, Weidong and Mao, Xiaojun},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-0534},
url = {https://mlanthology.org/neurips/2024/xia2024neurips-st/}
}