Optimizing Temperature for Language Models with Multi-Sample Inference

Abstract

Multi-sample aggregation strategies, such as majority voting and best-of-N sampling, are widely used in contemporary large language models (LLMs) to enhance predictive accuracy across various tasks. A key challenge in this process is temperature selection, which significantly impacts model performance. Existing approaches either rely on a fixed default temperature or require labeled validation data for tuning, which are often scarce and difficult to obtain. This paper addresses the challenge of automatically identifying the (near)-optimal temperature for different LLMs using multi-sample aggregation strategies, without relying on task-specific validation data. We provide a comprehensive analysis of temperature’s role in performance optimization, considering variations in model architectures, datasets, task types, model sizes, and predictive accuracy. Furthermore, we propose a novel entropy-based metric for automated temperature optimization, which consistently outperforms fixed-temperature baselines. Additionally, we incorporate a stochastic process model to enhance interpretability, offering deeper insights into the relationship between temperature and model performance.

Cite

Text

Du et al. "Optimizing Temperature for Language Models with Multi-Sample Inference." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Du et al. "Optimizing Temperature for Language Models with Multi-Sample Inference." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/du2025icml-optimizing/)

BibTeX

@inproceedings{du2025icml-optimizing,
  title     = {{Optimizing Temperature for Language Models with Multi-Sample Inference}},
  author    = {Du, Weihua and Yang, Yiming and Welleck, Sean},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {14648-14668},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/du2025icml-optimizing/}
}