The Differentiable Cross-Entropy Method

Abstract

We study the Cross-Entropy Method (CEM) for the non-convex optimization of a continuous and parameterized objective function and introduce a differentiable variant that enables us to differentiate the output of CEM with respect to the objective function’s parameters. In the machine learning setting this brings CEM inside of the end-to-end learning pipeline where this has otherwise been impossible. We show applications in a synthetic energy-based structured prediction task and in non-convex continuous control. In the control setting we show how to embed optimal action sequences into a lower-dimensional space. This enables us to use policy optimization to fine-tune modeling components by differentiating through the CEM-based controller.

Cite

Text

Amos and Yarats. "The Differentiable Cross-Entropy Method." International Conference on Machine Learning, 2020.

Markdown

[Amos and Yarats. "The Differentiable Cross-Entropy Method." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/amos2020icml-differentiable/)

BibTeX

@inproceedings{amos2020icml-differentiable,
  title     = {{The Differentiable Cross-Entropy Method}},
  author    = {Amos, Brandon and Yarats, Denis},
  booktitle = {International Conference on Machine Learning},
  year      = {2020},
  pages     = {291-302},
  volume    = {119},
  url       = {https://mlanthology.org/icml/2020/amos2020icml-differentiable/}
}