Improved Sample Complexity for Multiclass PAC Learning

Abstract

We aim to understand the optimal PAC sample complexity in multiclass learning. While finiteness of the Daniely-Shalev-Shwartz (DS) dimension has been shown to characterize the PAC learnability of a concept class [Brukhim, Carmon, Dinur, Moran, and Yehudayoff, 2022], there exist polylog factor gaps in the leading term of the sample complexity. In this paper, we reduce the gap in terms of the dependence on the error parameter to a single log factor and also propose two possible routes towards completely resolving the optimal sample complexity, each based on a key open question we formulate: one concerning list learning with bounded list size, the other concerning a new type of shifting for multiclass concept classes. We prove that a positive answer to either of the two questions would completely resolve the optimal sample complexity up to log factors of the DS dimension.

Cite

Text

Hanneke et al. "Improved Sample Complexity for Multiclass PAC Learning." Neural Information Processing Systems, 2024. doi:10.52202/079017-1356

Markdown

[Hanneke et al. "Improved Sample Complexity for Multiclass PAC Learning." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/hanneke2024neurips-improved/) doi:10.52202/079017-1356

BibTeX

@inproceedings{hanneke2024neurips-improved,
  title     = {{Improved Sample Complexity for Multiclass PAC Learning}},
  author    = {Hanneke, Steve and Moran, Shay and Zhang, Qian},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-1356},
  url       = {https://mlanthology.org/neurips/2024/hanneke2024neurips-improved/}
}