Piecewise Linear Parametrization of Policies: Towards Interpretable Deep Reinforcement Learning
Abstract
Learning inherently interpretable policies is a central challenge in the path to developing autonomous agents that humans can trust. We argue for the use of policies that are piecewise-linear. We carefully study to what extent they can retain the interpretable properties of linear policies while performing competitively with neural baselines. In particular, we propose the HyperCombinator (HC), a piecewise-linear neural architecture expressing a policy with a controllably small number of sub-policies. Each sub-policy is linear with respect to interpretable features, shedding light on the agent's decision process without needing an additional explanation model. We evaluate HC policies in control and navigation experiments, visualize the improved interpretability of the agent and highlight its trade-off with performance.
Cite
Text
Wabartha and Pineau. "Piecewise Linear Parametrization of Policies: Towards Interpretable Deep Reinforcement Learning." NeurIPS 2023 Workshops: XAIA, 2023.Markdown
[Wabartha and Pineau. "Piecewise Linear Parametrization of Policies: Towards Interpretable Deep Reinforcement Learning." NeurIPS 2023 Workshops: XAIA, 2023.](https://mlanthology.org/neuripsw/2023/wabartha2023neuripsw-piecewise/)BibTeX
@inproceedings{wabartha2023neuripsw-piecewise,
title = {{Piecewise Linear Parametrization of Policies: Towards Interpretable Deep Reinforcement Learning}},
author = {Wabartha, Maxime and Pineau, Joelle},
booktitle = {NeurIPS 2023 Workshops: XAIA},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/wabartha2023neuripsw-piecewise/}
}