MARS: Meaning-Aware Response Scoring for Uncertainty Estimation in Generative LLMs

Abstract

Generative Large Language Models (LLMs) are widely utilized for their excellence in various tasks. Eestimating the correctness of generative LLM outputs is an important task for enhanced reliability. Uncertainty Estimation (UE) in generative LLMs is an evolving domain, where SOTA probability-based methods commonly employ length-normalized scoring. In this work, we propose Meaning-Aware Response Scoring (MARS) as an alternative to length-normalized scoring for UE methods. MARS is a novel scoring function that considers the semantic contribution of each token in the generated sequence in the context of the question. We demonstrate that integrating MARS into UE methods results in a universal and significant improvement in UE performance. Code can be found \href{https://github.com/Ybakman/LLM_Uncertainty} here.

Cite

Text

Bakman et al. "MARS: Meaning-Aware Response Scoring  for Uncertainty Estimation in Generative LLMs." ICLR 2024 Workshops: SeT_LLM, 2024.

Markdown

[Bakman et al. "MARS: Meaning-Aware Response Scoring  for Uncertainty Estimation in Generative LLMs." ICLR 2024 Workshops: SeT_LLM, 2024.](https://mlanthology.org/iclrw/2024/bakman2024iclrw-mars/)

BibTeX

@inproceedings{bakman2024iclrw-mars,
  title     = {{MARS: Meaning-Aware Response Scoring  for Uncertainty Estimation in Generative LLMs}},
  author    = {Bakman, Yavuz Faruk and Yaldiz, Duygu Nur and Buyukates, Baturalp and Tao, Chenyang and Dimitriadis, Dimitrios and Avestimehr, Salman},
  booktitle = {ICLR 2024 Workshops: SeT_LLM},
  year      = {2024},
  url       = {https://mlanthology.org/iclrw/2024/bakman2024iclrw-mars/}
}