Sample Compression for Real-Valued Learners

Abstract

We give an algorithmically efficient version of the learner-to-compression scheme conversion in Moran and Yehudayoff (2016). We further extend this technique to real-valued hypotheses, to obtain a bounded-size sample compression scheme via an efficient reduction to a certain generic real-valued learning strategy. To our knowledge, this is the first general compressed regression result (regardless of efficiency or boundedness) guaranteeing uniform approximate reconstruction. Along the way, we develop a generic procedure for constructing weak real-valued learners out of abstract regressors; this result is also of independent interest. In particular, this result sheds new light on an open question of H. Simon (1997). We show applications to two regression problems: learning Lipschitz and bounded-variation functions.

Cite

Text

Hanneke et al. "Sample Compression for Real-Valued Learners." Proceedings of the 30th International Conference on Algorithmic Learning Theory, 2019.

Markdown

[Hanneke et al. "Sample Compression for Real-Valued Learners." Proceedings of the 30th International Conference on Algorithmic Learning Theory, 2019.](https://mlanthology.org/alt/2019/hanneke2019alt-sample/)

BibTeX

@inproceedings{hanneke2019alt-sample,
  title     = {{Sample Compression for Real-Valued Learners}},
  author    = {Hanneke, Steve and Kontorovich, Aryeh and Sadigurschi, Menachem},
  booktitle = {Proceedings of the 30th International Conference on Algorithmic Learning Theory},
  year      = {2019},
  pages     = {466-488},
  volume    = {98},
  url       = {https://mlanthology.org/alt/2019/hanneke2019alt-sample/}
}