Learning-to-Optimize with PAC-Bayesian Guarantees: Theoretical Considerations and Practical Implementation

Abstract

We use the PAC-Bayesian theory for the setting of learning-to-optimize. To the best of our knowledge, we present the first framework to learn optimization algorithms with provable generalization guarantees (PAC-Bayesian bounds) and explicit trade-off between convergence guarantees and convergence speed, which contrasts with the typical worst-case analysis. Our learned optimization algorithms provably outperform related ones derived from a worst-case analysis. The results rely on PAC-Bayesian bounds for general, possibly unbounded loss-functions based on exponential families. Further, we provide a concrete algorithmic realization of the framework and new methodologies for learning-to-optimize. Finally, we conduct four practically relevant experiments to support our theory. With this, we showcase that the provided learning framework yields optimization algorithms that provably outperform the state-of-the-art by orders of magnitude.

Cite

Text

Sucker et al. "Learning-to-Optimize with PAC-Bayesian Guarantees: Theoretical Considerations and Practical Implementation." Journal of Machine Learning Research, 2025.

Markdown

[Sucker et al. "Learning-to-Optimize with PAC-Bayesian Guarantees: Theoretical Considerations and Practical Implementation." Journal of Machine Learning Research, 2025.](https://mlanthology.org/jmlr/2025/sucker2025jmlr-learningtooptimize/)

BibTeX

@article{sucker2025jmlr-learningtooptimize,
  title     = {{Learning-to-Optimize with PAC-Bayesian Guarantees: Theoretical Considerations and Practical Implementation}},
  author    = {Sucker, Michael and Fadili, Jalal and Ochs, Peter},
  journal   = {Journal of Machine Learning Research},
  year      = {2025},
  pages     = {1-53},
  volume    = {26},
  url       = {https://mlanthology.org/jmlr/2025/sucker2025jmlr-learningtooptimize/}
}