Benchmarking and Improving Generator-Validator Consistency of Language Models
Abstract
As of September 2023, ChatGPT correctly answers “what is 7+8” with 15, but when asked “7+8=15, True or False” it responds with “False”. This inconsistency between generating and validating an answer is prevalent in language models (LMs) and erodes trust. In this paper, we propose a framework for measuring the consistency between generation and validation (which we call generator-validator consistency, or GV-consistency), finding that even GPT-4 (0613), a state-of-the-art LM, is GV-consistent only 76% of the time. To improve the consistency of LMs, we propose to finetune on the filtered generator and validator responses that are GV-consistent, and call this approach consistency fine-tuning. We find that this approach improves GV-consistency of Alpaca-30B from 60% to 93%, and the improvement extrapolates to unseen tasks and domains (e.g., GV-consistency for positive style transfers extrapolates to unseen styles like humor). In addition to improving consistency, consistency fine-tuning improves both generator quality and validator accuracy without using any labeled data. Evaluated across 6 tasks, including math questions, knowledge-intensive QA, and instruction following, our method improves generator quality by an average of 16% and validator accuracy by an average of 6.3% across all tasks.
Cite
Text
Li et al. "Benchmarking and Improving Generator-Validator Consistency of Language Models." International Conference on Learning Representations, 2024.Markdown
[Li et al. "Benchmarking and Improving Generator-Validator Consistency of Language Models." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/li2024iclr-benchmarking/)BibTeX
@inproceedings{li2024iclr-benchmarking,
title = {{Benchmarking and Improving Generator-Validator Consistency of Language Models}},
author = {Li, Xiang Lisa and Shrivastava, Vaishnavi and Li, Siyan and Hashimoto, Tatsunori and Liang, Percy},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/li2024iclr-benchmarking/}
}