Boss LLM: Adaptation via No-Regret Learning

Abstract

The diversity of Large Language Models (LLMs) calls for more effective strategies to combine their strengths across various tasks. In this work, we learn an adaptive mixture of multiple expert models as Boss LLM. By extending the multi-objective optimization with exponential weights (MOEW) algorithm, Boss LLM selects the most suitable model for a given prompt that could potentially span multiple categories with provable low regret for every category and expert model. Empirical results demonstrate that Boss LLM not only effectively adapts its mixture based upon the categories of a given prompt and improves upon the expert models, but also exhibits generalization properties.

Cite

Text

Feng et al. "Boss LLM: Adaptation via No-Regret Learning." ICLR 2025 Workshops: SSI-FM, 2025.

Markdown

[Feng et al. "Boss LLM: Adaptation via No-Regret Learning." ICLR 2025 Workshops: SSI-FM, 2025.](https://mlanthology.org/iclrw/2025/feng2025iclrw-boss/)

BibTeX

@inproceedings{feng2025iclrw-boss,
  title     = {{Boss LLM: Adaptation via No-Regret Learning}},
  author    = {Feng, Yu and Khare, Avishree and Nguyen, Nghia and Sengupta, Sikata Bela},
  booktitle = {ICLR 2025 Workshops: SSI-FM},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/feng2025iclrw-boss/}
}