Collaborative Discrete-Continuous Black-Box Prompt Learning for Language Models

Abstract

Large Scale Pre-Trained Language Models (PTMs) have demonstrated unprecedented capabilities across diverse natural language processing tasks. Adapting such models to downstream tasks is computationally intensive and time-consuming, particularly in black-box scenarios common in Language-Model-as-a-Service (LMaaS) environments, where model parameters and gradients are inaccessible. Recently, black-box prompt learning using zeroth-order gradients has emerged as a promising approach to address these challenges by optimizing learnable continuous prompts in embedding spaces, starting with \textit{randomly initialized discrete text prompts}. However, its reliance on randomly initialized discrete prompts limits adaptability to diverse downstream tasks or models. To address this limitation, this paper introduces ZO-PoG, a novel framework that optimizes prompts through a collaborative approach, combining Policy Gradient optimization for initial discrete text prompts and Zeroth-Order optimization for continuous prompts in embedding space. By optimizing collaboratively between discrete and continuous prompts, ZO-PoG maximizes adaptability to downstream tasks, achieving superior results without direct access to the model’s internal structures. Importantly, we establish the sub-linear convergence of ZO-PoG under mild assumptions. The experiments on different datasets demonstrate significant improvements in various tasks compared to the baselines.

Cite

Text

Zhang et al. "Collaborative Discrete-Continuous Black-Box Prompt Learning for Language Models." International Conference on Learning Representations, 2025.

Markdown

[Zhang et al. "Collaborative Discrete-Continuous Black-Box Prompt Learning for Language Models." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/zhang2025iclr-collaborative/)

BibTeX

@inproceedings{zhang2025iclr-collaborative,
  title     = {{Collaborative Discrete-Continuous Black-Box Prompt Learning for Language Models}},
  author    = {Zhang, Hualin and Zhang, Haozhen and Liu, Zhekai and Gu, Bin and Chang, Yi},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/zhang2025iclr-collaborative/}
}