Federated Stochastic Bilevel Optimization with Fully First-Order Gradients
Abstract
Federated stochastic bilevel optimization has been actively studied in recent years due to its widespread applications in machine learning. However, most existing federated stochastic bilevel optimization algorithms require the computation of second-order Hessian and Jacobian matrices, which leads to longer running times in practice. To address these challenges, we propose a novel federated stochastic variance-reduced bilevel gradient descent algorithm that relies solely on first-order oracles. Specifically, our approach does not require the computation of second-order Hessian and Jacobian matrices, significantly reducing running time. Furthermore, we introduce a novel learning rate mechanism, i.e., a constant single-time-scale learning rate, to coordinate the update of different variables. We also present a new strategy to establish the convergence rate of our algorithm. Finally, the extensive experimental results confirm the efficacy of our proposed algorithm.
Cite
Text
Zhang et al. "Federated Stochastic Bilevel Optimization with Fully First-Order Gradients." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/784Markdown
[Zhang et al. "Federated Stochastic Bilevel Optimization with Fully First-Order Gradients." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/zhang2025ijcai-federated/) doi:10.24963/IJCAI.2025/784BibTeX
@inproceedings{zhang2025ijcai-federated,
title = {{Federated Stochastic Bilevel Optimization with Fully First-Order Gradients}},
author = {Zhang, Yihan and Dhaipule, Rohit and Tan, Chiu C. and Ling, Haibin and Gao, Hongchang},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2025},
pages = {7047-7055},
doi = {10.24963/IJCAI.2025/784},
url = {https://mlanthology.org/ijcai/2025/zhang2025ijcai-federated/}
}