Efficient List-Decodable Regression Using Batches

Abstract

We demonstrate the use of batches in studying list-decodable linear regression, in which only $\alpha\in (0,1]$ fraction of batches contain genuine samples from a common distribution and the rest can contain arbitrary or even adversarial samples. When genuine batches have $\ge \tilde\Omega(1/\alpha)$ samples each, our algorithm can efficiently find a small list of potential regression parameters, with a high probability that one of them is close to the true parameter. This is the first polynomial time algorithm for list-decodable linear regression, and its sample complexity scales nearly linearly with the dimension of the covariates. The polynomial time algorithm is made possible by the batch structure and may not be feasible without it, as suggested by a recent Statistical Query lower bound (Diakonikolas et al., 2021b).

Cite

Text

Das et al. "Efficient List-Decodable Regression Using Batches." International Conference on Machine Learning, 2023.

Markdown

[Das et al. "Efficient List-Decodable Regression Using Batches." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/das2023icml-efficient/)

BibTeX

@inproceedings{das2023icml-efficient,
  title     = {{Efficient List-Decodable Regression Using Batches}},
  author    = {Das, Abhimanyu and Jain, Ayush and Kong, Weihao and Sen, Rajat},
  booktitle = {International Conference on Machine Learning},
  year      = {2023},
  pages     = {7025-7065},
  volume    = {202},
  url       = {https://mlanthology.org/icml/2023/das2023icml-efficient/}
}