Better than Random: Reliable NLG Human Evaluation with Constrained Active Sampling
Abstract
Human evaluation is viewed as a reliable evaluation method for NLG which is expensive and time-consuming. To save labor and costs, researchers usually perform human evaluation on a small subset of data sampled from the whole dataset in practice. However, different selection subsets will lead to different rankings of the systems. To give a more correct inter-system ranking and make the gold standard human evaluation more reliable, we propose a Constrained Active Sampling Framework (CASF) for reliable human judgment. CASF operates through a Learner, a Systematic Sampler and a Constrained Controller to select representative samples for getting a more correct inter-system ranking. Experiment results on 137 real NLG evaluation setups with 44 human evaluation metrics across 16 datasets and 5 NLG tasks demonstrate CASF receives 93.18\% top-ranked system recognition accuracy and ranks first or ranks second on 90.91\% of the human metrics with 0.83 overall inter-system ranking Kendall correlation. Code and data are publicly available online.
Cite
Text
Ruan et al. "Better than Random: Reliable NLG Human Evaluation with Constrained Active Sampling." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I17.29857Markdown
[Ruan et al. "Better than Random: Reliable NLG Human Evaluation with Constrained Active Sampling." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/ruan2024aaai-better/) doi:10.1609/AAAI.V38I17.29857BibTeX
@inproceedings{ruan2024aaai-better,
title = {{Better than Random: Reliable NLG Human Evaluation with Constrained Active Sampling}},
author = {Ruan, Jie and Pu, Xiao and Gao, Mingqi and Wan, Xiaojun and Zhu, Yuesheng},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2024},
pages = {18915-18923},
doi = {10.1609/AAAI.V38I17.29857},
url = {https://mlanthology.org/aaai/2024/ruan2024aaai-better/}
}