Consistent Adversarially Robust Linear Classification: Non-Parametric Setting

Abstract

For binary classification in $d$ dimensions, it is known that with a sample size of $n$, an excess adversarial risk of $O(d/n)$ is achievable under strong parametric assumptions about the underlying data distribution (e.g., assuming a Gaussian mixture model). In the case of well-separated distributions, this rate can be further refined to $O(1/n)$. Our work studies the non-parametric setting, where very little is known. With only mild regularity conditions on the conditional distribution of the features, we examine adversarial attacks with respect to arbitrary norms and introduce a straightforward yet effective estimator with provable consistency w.r.t adversarial risk. Our estimator is given by minimizing a series of smoothed versions of the robust 0/1 loss, with a smoothing bandwidth that adapts to both $n$ and $d$. Furthermore, we demonstrate that our estimator can achieve the minimax excess adversarial risk of $\widetilde O(\sqrt{d/n})$ for linear classifiers, at the cost of solving possibly rougher optimization problems.

Cite

Text

Dohmatob. "Consistent Adversarially Robust Linear Classification: Non-Parametric Setting." International Conference on Machine Learning, 2024.

Markdown

[Dohmatob. "Consistent Adversarially Robust Linear Classification: Non-Parametric Setting." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/dohmatob2024icml-consistent/)

BibTeX

@inproceedings{dohmatob2024icml-consistent,
  title     = {{Consistent Adversarially Robust Linear Classification: Non-Parametric Setting}},
  author    = {Dohmatob, Elvis},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {11149-11164},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/dohmatob2024icml-consistent/}
}