Learning in the Limit with Adversarial Disturbances

Abstract

We study distribution-dependent, data-dependent, learning in the limit with adversarial disturbance. We consider an optimization-based approach to learning binary classifiers from data under worst-case assumptions on the disturbance. The learning process is modeled as a decision-maker who seeks to minimize generalization error, given access only to possibly maliciously corrupted data. Two models for the nature of the disturbance are considered: disturbance in the labels of a certain fraction of the data, and disturbance that also affects the position of the data points. We provide distributiondependent bounds on the amount of error as a function of the noise level for the two models, and describe the optimal strategy of the decision-maker, as well as the worst-case disturbance.

Cite

Text

Caramanis and Mannor. "Learning in the Limit with Adversarial Disturbances." Annual Conference on Computational Learning Theory, 2008.

Markdown

[Caramanis and Mannor. "Learning in the Limit with Adversarial Disturbances." Annual Conference on Computational Learning Theory, 2008.](https://mlanthology.org/colt/2008/caramanis2008colt-learning/)

BibTeX

@inproceedings{caramanis2008colt-learning,
  title     = {{Learning in the Limit with Adversarial Disturbances}},
  author    = {Caramanis, Constantine and Mannor, Shie},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2008},
  pages     = {467-478},
  url       = {https://mlanthology.org/colt/2008/caramanis2008colt-learning/}
}