Hessian Free Efficient Single Loop Iterative Differentiation Methods for Bi-Level Optimization Problems

Abstract

Bilevel optimization problems have been actively studied in recent machine learning research due to their broad applications. In this work, we investigate single-loop methods with iterative differentiation (ITD) for nonconvex bilevel optimization problems. For deterministic bilevel problems, we propose an efficient single-loop ITD-type method (ES-ITDM). Our method employs historical updates to approximate the hypergradient. More importantly, based on ES-ITDM, we propose a new method that avoids computing Hessians. This Hessian-free method requires fewer backpropagations and thus has a lower computational cost. We analyze the convergence properties of the proposed methods in two aspects. We provide the convergence rates of the sequences generated by ES-ITD based on the Kurdyka-\L ojasiewicz (KL) property. We also show that the Hessian-free stochastic ES-ITDM has the best-known complexity while has cheaper computation. The empirical studies show that our Hessian-free stochastic variant is more efficient than existing Hessian-free methods and other state-of-the-art bilevel optimization approaches.

Cite

Text

Yu et al. "Hessian Free Efficient Single Loop Iterative Differentiation Methods for Bi-Level Optimization Problems." Transactions on Machine Learning Research, 2024.

Markdown

[Yu et al. "Hessian Free Efficient Single Loop Iterative Differentiation Methods for Bi-Level Optimization Problems." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/yu2024tmlr-hessian/)

BibTeX

@article{yu2024tmlr-hessian,
  title     = {{Hessian Free Efficient Single Loop Iterative Differentiation Methods for Bi-Level Optimization Problems}},
  author    = {Yu, Peiran and Li, Junyi and Huang, Heng},
  journal   = {Transactions on Machine Learning Research},
  year      = {2024},
  url       = {https://mlanthology.org/tmlr/2024/yu2024tmlr-hessian/}
}