Scalable Gradients for Stochastic Differential Equations

Abstract

The adjoint sensitivity method scalably computes gradients of solutions to ordinary differential equations. We generalize this method to stochastic differential equations, allowing time-efficient and constant-memory computation of gradients with high-order adaptive solvers. Specifically, we derive a stochastic differentialequation whose solution is the gradient, a memory-efficient algorithm for cachingnoise, and conditions under which numerical solutions converge. In addition, we combine our method with gradient-based stochastic variational inference for latent stochastic differential equations. We use our method to fit stochastic dynamics defined by neural networks, achieving competitive performance ona 50-dimensional motion capture dataset.

Cite

Text

Li et al. "Scalable Gradients for Stochastic Differential Equations." Artificial Intelligence and Statistics, 2020.

Markdown

[Li et al. "Scalable Gradients for Stochastic Differential Equations." Artificial Intelligence and Statistics, 2020.](https://mlanthology.org/aistats/2020/li2020aistats-scalable/)

BibTeX

@inproceedings{li2020aistats-scalable,
  title     = {{Scalable Gradients for Stochastic Differential Equations}},
  author    = {Li, Xuechen and Wong, Ting-Kam Leonard and Chen, Ricky T. Q. and Duvenaud, David},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2020},
  pages     = {3870-3882},
  volume    = {108},
  url       = {https://mlanthology.org/aistats/2020/li2020aistats-scalable/}
}