Stable Adaptive Control with Online Learning

Abstract

Learning algorithms have enjoyed numerous successes in robotic control tasks. In problems with time-varying dynamics, online learning methods have also proved to be a powerful tool for automatically tracking and/or adapting to the changing circumstances. However, for safety-critical ap- plications such as airplane flight, the adoption of these algorithms has been significantly hampered by their lack of safety, such as "stability," guarantees. Rather than trying to show difficult, a priori, stability guar- antees for specific learning methods, in this paper we propose a method for "monitoring" the controllers suggested by the learning algorithm on- line, and rejecting controllers leading to instability. We prove that even if an arbitrary online learning method is used with our algorithm to control a linear dynamical system, the resulting system is stable.

Cite

Text

Kim and Ng. "Stable Adaptive Control with Online Learning." Neural Information Processing Systems, 2004.

Markdown

[Kim and Ng. "Stable Adaptive Control with Online Learning." Neural Information Processing Systems, 2004.](https://mlanthology.org/neurips/2004/kim2004neurips-stable/)

BibTeX

@inproceedings{kim2004neurips-stable,
  title     = {{Stable Adaptive Control with Online Learning}},
  author    = {Kim, H. J. and Ng, Andrew Y.},
  booktitle = {Neural Information Processing Systems},
  year      = {2004},
  pages     = {977-984},
  url       = {https://mlanthology.org/neurips/2004/kim2004neurips-stable/}
}