Formalizing the Problem of Side Effect Regularization
Abstract
AI objectives are often hard to specify properly. Some approaches tackle this problem by regularizing the AI’s side effects: Agents must weigh off “how much of a mess they make” with an imperfectly specified proxy objective. We propose a formal criterion for side effect regularization via the assistance game framework [Shah et al., 2021]. In these games, the agent solves a partially observable Markov decision process (POMDP) representing its uncertainty about the objective function it should optimize. We consider the setting where the true objective is revealed to the agent at a later time step. We show that this POMDP is solved by trading off the proxy reward with the agent’s ability to achieve a range of future tasks. We empirically demonstrate the reasonableness of our problem formalization via ground-truth evaluation in two gridworld environments.
Cite
Text
Turner et al. "Formalizing the Problem of Side Effect Regularization." NeurIPS 2022 Workshops: MLSW, 2022.Markdown
[Turner et al. "Formalizing the Problem of Side Effect Regularization." NeurIPS 2022 Workshops: MLSW, 2022.](https://mlanthology.org/neuripsw/2022/turner2022neuripsw-formalizing/)BibTeX
@inproceedings{turner2022neuripsw-formalizing,
title = {{Formalizing the Problem of Side Effect Regularization}},
author = {Turner, Alexander Matt and Saxena, Aseem and Tadepalli, Prasad},
booktitle = {NeurIPS 2022 Workshops: MLSW},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/turner2022neuripsw-formalizing/}
}