Generalizing While Preserving Monotonicity in Comparison-Based Preference Learning Models
Abstract
If you tell a learning model that you prefer an alternative $a$ over another alternative $b$, then you probably expect the model to be *monotone*, that is, the valuation of $a$ increases, and that of $b$ decreases. Yet, perhaps surprisingly, many widely deployed comparison-based preference learning models, including large language models, fail to have this guarantee. Until now, the only comparison-based preference learning algorithms that were proved to be monotone are the Generalized Bradley-Terry models. Yet, these models are unable to generalize to uncompared data. In this paper, we advance the understanding of the set of models with generalization ability that are *monotone*. Namely, we propose a new class of Linear Generalized Bradley-Terry models with Diffusion Priors, and identify sufficient conditions on alternatives' embeddings that guarantee monotonicity. Our experiments show that this monotonicity is far from being a general guarantee, and that our new class of generalizing models improves accuracy, especially when the dataset is limited.
Cite
Text
Fageot et al. "Generalizing While Preserving Monotonicity in Comparison-Based Preference Learning Models." Advances in Neural Information Processing Systems, 2025.Markdown
[Fageot et al. "Generalizing While Preserving Monotonicity in Comparison-Based Preference Learning Models." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/fageot2025neurips-generalizing/)BibTeX
@inproceedings{fageot2025neurips-generalizing,
title = {{Generalizing While Preserving Monotonicity in Comparison-Based Preference Learning Models}},
author = {Fageot, Julien and Blanchard, Peva and Bareilles, Gilles and Hoang, Lê-Nguyên},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/fageot2025neurips-generalizing/}
}