First-Order Methods for Linearly Constrained Bilevel Optimization

Abstract

Algorithms for bilevel optimization often encounter Hessian computations, which are prohibitive in high dimensions. While recent works offer first-order methods for unconstrained bilevel problems, the constrained setting remains relatively underexplored. We present first-order linearly constrained optimization methods with finite-time hypergradient stationarity guarantees. For linear equality constraints, we attain $\epsilon$-stationarity in $\widetilde{O}(\epsilon^{-2})$ gradient oracle calls, which is nearly-optimal. For linear inequality constraints, we attain $(\delta,\epsilon)$-Goldstein stationarity in $\widetilde{O}(d{\delta^{-1} \epsilon^{-3}})$ gradient oracle calls, where $d$ is the upper-level dimension. Finally, we obtain for the linear inequality setting dimension-free rates of $\widetilde{O}({\delta^{-1} \epsilon^{-4}})$ oracle complexity under the additional assumption of oracle access to the optimal dual variable. Along the way, we develop new nonsmooth nonconvex optimization methods with inexact oracles. Our numerical experiments verify these guarantees.

Cite

Text

Kornowski et al. "First-Order Methods for Linearly Constrained Bilevel Optimization." Neural Information Processing Systems, 2024. doi:10.52202/079017-4491

Markdown

[Kornowski et al. "First-Order Methods for Linearly Constrained Bilevel Optimization." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/kornowski2024neurips-firstorder/) doi:10.52202/079017-4491

BibTeX

@inproceedings{kornowski2024neurips-firstorder,
  title     = {{First-Order Methods for Linearly Constrained Bilevel Optimization}},
  author    = {Kornowski, Guy and Padmanabhan, Swati and Wang, Kai and Zhang, Jimmy and Sra, Suvrit},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-4491},
  url       = {https://mlanthology.org/neurips/2024/kornowski2024neurips-firstorder/}
}