Large-Scale Learning with Less RAM via Randomization
Abstract
We reduce the memory footprint of popular large-scale online learning methods by projecting our weight vector onto a coarse discrete set using randomized rounding. Compared to standard 32-bit float encodings, this reduces RAM usage by more than 50% during training and by up 95% when making predictions from a fixed model, with almost no loss in accuracy. We also show that randomized counting can be used to implement per-coordinate learning rates, improving model quality with little additional RAM. We prove these memory-saving methods achieve regret guarantees similar to their exact variants. Empirical evaluation confirms excellent performance, dominating standard approaches across memory versus accuracy tradeoffs.
Cite
Text
Golovin et al. "Large-Scale Learning with Less RAM via Randomization." International Conference on Machine Learning, 2013.Markdown
[Golovin et al. "Large-Scale Learning with Less RAM via Randomization." International Conference on Machine Learning, 2013.](https://mlanthology.org/icml/2013/golovin2013icml-largescale/)BibTeX
@inproceedings{golovin2013icml-largescale,
title = {{Large-Scale Learning with Less RAM via Randomization}},
author = {Golovin, Daniel and Sculley, D. and McMahan, Brendan and Young, Michael},
booktitle = {International Conference on Machine Learning},
year = {2013},
pages = {325-333},
volume = {28},
url = {https://mlanthology.org/icml/2013/golovin2013icml-largescale/}
}