On-Line Learning with Malicious Noise and the Closure Algorithm

Abstract

We investigate a variant of the on-line learning model for classes of 0,l-valued functions (concepts) in which the labels of a certain amount of the input instances are corrupted by adversarial noise. We propose an extension of a general learning strategy, known as “Closure Algorithm”, to this noise model, and show a worst-case mistake bound of m +( d +1) K for learning an arbitrary intersection-closed concept class C , where K is the number of noisy labels, d is a combinatorial parameter measuring C 's complexity, and m is the worst-case mistake bound of the Closure Algorithm for learning C in the noise-free model. For several concept classes our extended Closure Algorithm is efficient and can tolerate a noise rate equal to the information-theoretic upper bound. We also show how to efficiently turn any algorithm for the on-line noise model into a learning algorithm for the PAC model with malicious noise.

Cite

Text

Auer and Cesa-Bianchi. "On-Line Learning with Malicious Noise and the Closure Algorithm." International Conference on Algorithmic Learning Theory, 1994. doi:10.1007/3-540-58520-6_67

Markdown

[Auer and Cesa-Bianchi. "On-Line Learning with Malicious Noise and the Closure Algorithm." International Conference on Algorithmic Learning Theory, 1994.](https://mlanthology.org/alt/1994/auer1994alt-online/) doi:10.1007/3-540-58520-6_67

BibTeX

@inproceedings{auer1994alt-online,
  title     = {{On-Line Learning with Malicious Noise and the Closure Algorithm}},
  author    = {Auer, Peter and Cesa-Bianchi, Nicolò},
  booktitle = {International Conference on Algorithmic Learning Theory},
  year      = {1994},
  pages     = {229-247},
  doi       = {10.1007/3-540-58520-6_67},
  url       = {https://mlanthology.org/alt/1994/auer1994alt-online/}
}