Attribute Efficient Linear Regression with Distribution-Dependent Sampling

Abstract

We consider a budgeted learning setting, where the learner can only choose and observe a small subset of the attributes of each training example. We develop efficient algorithms for Ridge and Lasso linear regression, which utilize the geometry of the data by a novel distribution-dependent sampling scheme, and have excess risk bounds which are better a factor of up to O(d/k) over the state-of-the-art, where d is the dimension and k+1 is the number of observed attributes per example. Moreover, under reasonable assumptions, our algorithms are the first in our setting which can provably use *less* attributes than full-information algorithms, which is the main concern in budgeted learning. We complement our theoretical analysis with experiments which support our claims.

Cite

Text

Kukliansky and Shamir. "Attribute Efficient Linear Regression with Distribution-Dependent Sampling." International Conference on Machine Learning, 2015.

Markdown

[Kukliansky and Shamir. "Attribute Efficient Linear Regression with Distribution-Dependent Sampling." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/kukliansky2015icml-attribute/)

BibTeX

@inproceedings{kukliansky2015icml-attribute,
  title     = {{Attribute Efficient Linear Regression with Distribution-Dependent Sampling}},
  author    = {Kukliansky, Doron and Shamir, Ohad},
  booktitle = {International Conference on Machine Learning},
  year      = {2015},
  pages     = {153-161},
  volume    = {37},
  url       = {https://mlanthology.org/icml/2015/kukliansky2015icml-attribute/}
}