On Overcoming Miscalibrated Conversational Priors in LLM-Based ChatBots

Abstract

We explore the use of Large Language Model (LLM-based) chatbots to power recommender systems. We observe that the chatbots respond poorly when they encounter under-specified requests (e.g., they make incorrect assumptions, hedge with a long response, or refuse to answer). We conjecture that such miscalibrated response tendencies (i.e., conversational priors) can be attributed to LLM fine-tuning by annotators — single-turn annotations may not capture multi-turn conversation utility, and the annotators’ preferences may not even be representative of users interacting with a recommender system. We first analyze public LLM chat logs to conclude that query under-specification is common. Next, we study synthetic recommendation problems with known but latent item utilities, and frame them as Partially Observed Decision Processes (PODP). We find that pre-trained LLMs can be sub-optimal for PODPs and derive better policies that clarify under-specified queries when appropriate. Then, we re-calibrate LLMs by prompting them with learned control messages to approximate the improved policy. Finally, we show empirically that our lightweight learning approach effectively uses logged conversation data to re-calibrate the response strategies of LLM-based chatbots for recommendation tasks.

Cite

Text

Herlihy et al. "On Overcoming Miscalibrated Conversational Priors in LLM-Based ChatBots." Uncertainty in Artificial Intelligence, 2024.

Markdown

[Herlihy et al. "On Overcoming Miscalibrated Conversational Priors in LLM-Based ChatBots." Uncertainty in Artificial Intelligence, 2024.](https://mlanthology.org/uai/2024/herlihy2024uai-overcoming/)

BibTeX

@inproceedings{herlihy2024uai-overcoming,
  title     = {{On Overcoming Miscalibrated Conversational Priors in LLM-Based ChatBots}},
  author    = {Herlihy, Christine and Neville, Jennifer and Schnabel, Tobias and Swaminathan, Adith},
  booktitle = {Uncertainty in Artificial Intelligence},
  year      = {2024},
  pages     = {1599-1620},
  volume    = {244},
  url       = {https://mlanthology.org/uai/2024/herlihy2024uai-overcoming/}
}