Efficient Gradient Approximation Method for Constrained Bilevel Optimization
Abstract
Bilevel optimization has been developed for many machine learning tasks with large-scale and high-dimensional data. This paper considers a constrained bilevel optimization problem, where the lower-level optimization problem is convex with equality and inequality constraints and the upper-level optimization problem is non-convex. The overall objective function is non-convex and non-differentiable. To solve the problem, we develop a gradient-based approach, called gradient approximation method, which determines the descent direction by computing several representative gradients of the objective function inside a neighborhood of the current estimate. We show that the algorithm asymptotically converges to the set of Clarke stationary points, and demonstrate the efficacy of the algorithm by the experiments on hyperparameter optimization and meta-learning.
Cite
Text
Xu and Zhu. "Efficient Gradient Approximation Method for Constrained Bilevel Optimization." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I10.26473Markdown
[Xu and Zhu. "Efficient Gradient Approximation Method for Constrained Bilevel Optimization." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/xu2023aaai-efficient-a/) doi:10.1609/AAAI.V37I10.26473BibTeX
@inproceedings{xu2023aaai-efficient-a,
title = {{Efficient Gradient Approximation Method for Constrained Bilevel Optimization}},
author = {Xu, Siyuan and Zhu, Minghui},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {12509-12517},
doi = {10.1609/AAAI.V37I10.26473},
url = {https://mlanthology.org/aaai/2023/xu2023aaai-efficient-a/}
}