Policy Aggregation
Abstract
We consider the challenge of AI value alignment with multiple individuals that have different reward functions and optimal policies in an underlying Markov decision process. We formalize this problem as one of *policy aggregation*, where the goal is to identify a desirable collective policy. We argue that an approach informed by social choice theory is especially suitable. Our key insight is that social choice methods can be reinterpreted by identifying ordinal preferences with volumes of subsets of the *state-action occupancy polytope*. Building on this insight, we demonstrate that a variety of methods — including approval voting, Borda count, the proportional veto core, and quantile fairness — can be practically applied to policy aggregation.
Cite
Text
Alamdari et al. "Policy Aggregation." NeurIPS 2024 Workshops: Pluralistic-Alignment, 2024.Markdown
[Alamdari et al. "Policy Aggregation." NeurIPS 2024 Workshops: Pluralistic-Alignment, 2024.](https://mlanthology.org/neuripsw/2024/alamdari2024neuripsw-policy/)BibTeX
@inproceedings{alamdari2024neuripsw-policy,
title = {{Policy Aggregation}},
author = {Alamdari, Parand A. and Ebadian, Soroush and Procaccia, Ariel D.},
booktitle = {NeurIPS 2024 Workshops: Pluralistic-Alignment},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/alamdari2024neuripsw-policy/}
}