Private Fine-Tuning of Large Language Models with Zeroth-Order Optimization

Abstract

Differentially private stochastic gradient descent (DP-SGD) allows models to be trained in a privacy-preserving manner, but has proven difficult to scale to the era of foundation models. We introduce DP-ZO, a private fine-tuning method for large language models by privatizing zeroth order optimization methods. A key insight into the design of our method is that the direction of the gradient in the zeroth-order optimization we use is random and the only information from training data is the step size, i.e., a scalar. Therefore, we only need to privatize the scalar step size, which is memory-efficient. DP-ZO provides a strong privacy-utility trade-off across different tasks, and model sizes that are comparable to DP-SGD in $(\varepsilon,\delta)$-DP. Notably, DP-ZO possesses significant advantages over DP-SGD in memory efficiency, and obtains higher utility in $\varepsilon$-DP when using the Laplace mechanism.

Cite

Text

Tang et al. "Private Fine-Tuning of Large Language Models with Zeroth-Order Optimization." Transactions on Machine Learning Research, 2025.

Markdown

[Tang et al. "Private Fine-Tuning of Large Language Models with Zeroth-Order Optimization." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/tang2025tmlr-private/)

BibTeX

@article{tang2025tmlr-private,
  title     = {{Private Fine-Tuning of Large Language Models with Zeroth-Order Optimization}},
  author    = {Tang, Xinyu and Panda, Ashwinee and Nasr, Milad and Mahloujifar, Saeed and Mittal, Prateek},
  journal   = {Transactions on Machine Learning Research},
  year      = {2025},
  url       = {https://mlanthology.org/tmlr/2025/tang2025tmlr-private/}
}