Accelerating Quadratic Optimization with Reinforcement Learning

Abstract

First-order methods for quadratic optimization such as OSQP are widely used for large-scale machine learning and embedded optimal control, where many related problems must be rapidly solved. These methods face two persistent challenges: manual hyperparameter tuning and convergence time to high-accuracy solutions. To address these, we explore how Reinforcement Learning (RL) can learn a policy to tune parameters to accelerate convergence. In experiments with well-known QP benchmarks we find that our RL policy, RLQP, significantly outperforms state-of-the-art QP solvers by up to 3x. RLQP generalizes surprisingly well to previously unseen problems with varying dimension and structure from different applications, including the QPLIB, Netlib LP and Maros-M{\'e}sz{\'a}ros problems. Code, models, and videos are available at https://berkeleyautomation.github.io/rlqp/.

Cite

Text

Ichnowski et al. "Accelerating Quadratic Optimization with Reinforcement Learning." Neural Information Processing Systems, 2021.

Markdown

[Ichnowski et al. "Accelerating Quadratic Optimization with Reinforcement Learning." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/ichnowski2021neurips-accelerating/)

BibTeX

@inproceedings{ichnowski2021neurips-accelerating,
  title     = {{Accelerating Quadratic Optimization with Reinforcement Learning}},
  author    = {Ichnowski, Jeffrey and Jain, Paras and Stellato, Bartolomeo and Banjac, Goran and Luo, Michael and Borrelli, Francesco and Gonzalez, Joseph E and Stoica, Ion and Goldberg, Ken},
  booktitle = {Neural Information Processing Systems},
  year      = {2021},
  url       = {https://mlanthology.org/neurips/2021/ichnowski2021neurips-accelerating/}
}