ST$_k$: A Scalable Module for Solving Top-K Problems

Abstract

The cost of ranking becomes significant in the new stage of deep learning. We propose ST$_k$, a fully differentiable module with a single trainable parameter, designed to solve the Top-k problem without requiring additional time or GPU memory. Due to its fully differentiable nature, ST$_k$ can be embedded end-to-end into neural networks and optimize the Top-k problems within a unified computational graph. We apply ST$_k$ to the Average Top-k Loss (AT$_k$), which inherently faces a Top-k problem. The proposed ST$_k$ Loss outperforms AT$_k$ Loss and achieves the best average performance on multiple benchmarks, with the lowest standard deviation. With the assistance of ST$_k$ Loss, we surpass the state-of-the-art (SOTA) on both CIFAR-100-LT and Places-LT leaderboards.

Cite

Text

Xia et al. "ST$_k$: A Scalable Module for Solving Top-K Problems." Neural Information Processing Systems, 2024. doi:10.52202/079017-0534

Markdown

[Xia et al. "ST$_k$: A Scalable Module for Solving Top-K Problems." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/xia2024neurips-st/) doi:10.52202/079017-0534

BibTeX

@inproceedings{xia2024neurips-st,
  title     = {{ST$_k$: A Scalable Module for Solving Top-K Problems}},
  author    = {Xia, Hanchen and Liu, Weidong and Mao, Xiaojun},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-0534},
  url       = {https://mlanthology.org/neurips/2024/xia2024neurips-st/}
}