The Differentiable Cross-Entropy Method
Abstract
We study the Cross-Entropy Method (CEM) for the non-convex optimization of a continuous and parameterized objective function and introduce a differentiable variant that enables us to differentiate the output of CEM with respect to the objective function’s parameters. In the machine learning setting this brings CEM inside of the end-to-end learning pipeline where this has otherwise been impossible. We show applications in a synthetic energy-based structured prediction task and in non-convex continuous control. In the control setting we show how to embed optimal action sequences into a lower-dimensional space. This enables us to use policy optimization to fine-tune modeling components by differentiating through the CEM-based controller.
Cite
Text
Amos and Yarats. "The Differentiable Cross-Entropy Method." International Conference on Machine Learning, 2020.Markdown
[Amos and Yarats. "The Differentiable Cross-Entropy Method." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/amos2020icml-differentiable/)BibTeX
@inproceedings{amos2020icml-differentiable,
title = {{The Differentiable Cross-Entropy Method}},
author = {Amos, Brandon and Yarats, Denis},
booktitle = {International Conference on Machine Learning},
year = {2020},
pages = {291-302},
volume = {119},
url = {https://mlanthology.org/icml/2020/amos2020icml-differentiable/}
}