Learning from Personal Preferences

Abstract

Machine learning practitioners frequently use majority vote to resolve disagreement in multi-annotator datasets. While this approach is natural in settings where a single ground truth label exists for each instance, it hides the presence of disagreement for subjective annotation tasks. In domains such as language modeling, information retrieval, and top-k recommendation, models must avoid suppressing minority views and express when the answer to a query is contentious. We propose personalized error metrics to formalize the requirement of strong performance across a heterogeneous user population. Following this framework, we develop an algorithm for training an ensemble of models, each specialized for a different segment of the population.

Cite

Text

Jiang et al. "Learning from Personal Preferences." NeurIPS 2024 Workshops: Pluralistic-Alignment, 2024.

Markdown

[Jiang et al. "Learning from Personal Preferences." NeurIPS 2024 Workshops: Pluralistic-Alignment, 2024.](https://mlanthology.org/neuripsw/2024/jiang2024neuripsw-learning/)

BibTeX

@inproceedings{jiang2024neuripsw-learning,
  title     = {{Learning from Personal Preferences}},
  author    = {Jiang, Kelly and Ustun, Berk and Hullman, Jessica},
  booktitle = {NeurIPS 2024 Workshops: Pluralistic-Alignment},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/jiang2024neuripsw-learning/}
}