Bandit Multiclass Linear Classification: Efficient Algorithms for the Separable Case

Abstract

We study the problem of efficient online multiclass linear classification with bandit feedback, where all examples belong to one of $K$ classes and lie in the $d$-dimensional Euclidean space. Previous works have left open the challenge of designing efficient algorithms with finite mistake bounds when the data is linearly separable by a margin $\gamma$. In this work, we take a first step towards this problem. We consider two notions of linear separability: strong and weak. 1. Under the strong linear separability condition, we design an efficient algorithm that achieves a near-optimal mistake bound of $O\left(\frac{K}{\gamma^2} \right)$. 2. Under the more challenging weak linear separability condition, we design an efficient algorithm with a mistake bound of $2^{\widetilde{O}(\min(K \log^2 \frac{1}{\gamma}, \sqrt{\frac{1}{\gamma}} \log K))}$. Our algorithm is based on kernel Perceptron, which is inspired by the work of Klivans & Servedio (2008) on improperly learning intersection of halfspaces.

Cite

Text

Beygelzimer et al. "Bandit Multiclass Linear Classification: Efficient Algorithms for the Separable Case." International Conference on Machine Learning, 2019.

Markdown

[Beygelzimer et al. "Bandit Multiclass Linear Classification: Efficient Algorithms for the Separable Case." International Conference on Machine Learning, 2019.](https://mlanthology.org/icml/2019/beygelzimer2019icml-bandit/)

BibTeX

@inproceedings{beygelzimer2019icml-bandit,
  title     = {{Bandit Multiclass Linear Classification: Efficient Algorithms for the Separable Case}},
  author    = {Beygelzimer, Alina and Pal, David and Szorenyi, Balazs and Thiruvenkatachari, Devanathan and Wei, Chen-Yu and Zhang, Chicheng},
  booktitle = {International Conference on Machine Learning},
  year      = {2019},
  pages     = {624-633},
  volume    = {97},
  url       = {https://mlanthology.org/icml/2019/beygelzimer2019icml-bandit/}
}