Large-Scale Learning with Less RAM via Randomization

Abstract

We reduce the memory footprint of popular large-scale online learning methods by projecting our weight vector onto a coarse discrete set using randomized rounding. Compared to standard 32-bit float encodings, this reduces RAM usage by more than 50% during training and by up 95% when making predictions from a fixed model, with almost no loss in accuracy. We also show that randomized counting can be used to implement per-coordinate learning rates, improving model quality with little additional RAM. We prove these memory-saving methods achieve regret guarantees similar to their exact variants. Empirical evaluation confirms excellent performance, dominating standard approaches across memory versus accuracy tradeoffs.

Cite

Text

Golovin et al. "Large-Scale Learning with Less RAM via Randomization." International Conference on Machine Learning, 2013.

Markdown

[Golovin et al. "Large-Scale Learning with Less RAM via Randomization." International Conference on Machine Learning, 2013.](https://mlanthology.org/icml/2013/golovin2013icml-largescale/)

BibTeX

@inproceedings{golovin2013icml-largescale,
  title     = {{Large-Scale Learning with Less RAM via Randomization}},
  author    = {Golovin, Daniel and Sculley, D. and McMahan, Brendan and Young, Michael},
  booktitle = {International Conference on Machine Learning},
  year      = {2013},
  pages     = {325-333},
  volume    = {28},
  url       = {https://mlanthology.org/icml/2013/golovin2013icml-largescale/}
}