Learning from Personal Preferences
Abstract
Machine learning practitioners frequently use majority vote to resolve disagreement in multi-annotator datasets. While this approach is natural in settings where a single ground truth label exists for each instance, it hides the presence of disagreement for subjective annotation tasks. In domains such as language modeling, information retrieval, and top-k recommendation, models must avoid suppressing minority views and express when the answer to a query is contentious. We propose personalized error metrics to formalize the requirement of strong performance across a heterogeneous user population. Following this framework, we develop an algorithm for training an ensemble of models, each specialized for a different segment of the population.
Cite
Text
Jiang et al. "Learning from Personal Preferences." NeurIPS 2024 Workshops: Pluralistic-Alignment, 2024.Markdown
[Jiang et al. "Learning from Personal Preferences." NeurIPS 2024 Workshops: Pluralistic-Alignment, 2024.](https://mlanthology.org/neuripsw/2024/jiang2024neuripsw-learning/)BibTeX
@inproceedings{jiang2024neuripsw-learning,
title = {{Learning from Personal Preferences}},
author = {Jiang, Kelly and Ustun, Berk and Hullman, Jessica},
booktitle = {NeurIPS 2024 Workshops: Pluralistic-Alignment},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/jiang2024neuripsw-learning/}
}