Iterative Foundation Model Fine-Tuning on Multiple Rewards

Abstract

Fine-tuning foundation models has emerged as a powerful approach for generating objects with specific desired properties. Reinforcement learning (RL) provides an effective framework for this purpose, enabling models to generate outputs that maximize a given reward function. However, in many applications such as text generation and drug discovery, it can be suboptimal to optimize using a single reward signal, as multiple evaluation criteria are often necessary. This paper proposes a novel reinforcement learning-based method for fine-tuning foundation models using multiple reward signals. By employing an iterative fine-tuning strategy across these rewards, our approach generalizes state-of-the-art RL-based methods. We further provide a theoretical analysis that offers insights into the performance of multi-reward RL fine-tuning. Experimental results across diverse domains including text, biological sequence, and small molecule generation, demonstrate the effectiveness of the proposed algorithm compared to state-of-the-art baselines.

Cite

Text

Ghari et al. "Iterative Foundation Model Fine-Tuning on Multiple Rewards." Advances in Neural Information Processing Systems, 2025.

Markdown

[Ghari et al. "Iterative Foundation Model Fine-Tuning on Multiple Rewards." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/ghari2025neurips-iterative/)

BibTeX

@inproceedings{ghari2025neurips-iterative,
  title     = {{Iterative Foundation Model Fine-Tuning on Multiple Rewards}},
  author    = {Ghari, Pouya M. and Sciabola, Simone and Wang, Ye},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/ghari2025neurips-iterative/}
}