Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games

Abstract

Nowadays, AI-based techniques, such as deep neural networks (DNNs), are widely deployed in autonomous systems for complex mission requirements (e.g., motion planning in robotics). However, DNNs-based controllers are typically very complex, and it is very hard to formally verify their correctness, potentially causing severe risks for safety-critical autonomous systems. In this paper, we propose a construction scheme for a so-called Safe-visor architecture to sandbox DNNs-based controllers. Particularly, we consider the construction under a stochastic game framework to provide a system-level safety guarantee which is robust to noises and disturbances. A supervisor is built to check the control inputs provided by a DNNs-based controller and decide whether to accept them. Meanwhile, a safety advisor is running in parallel to provide fallback control inputs in case the DNN-based controller is rejected. We demonstrate the proposed approaches on a quadrotor employing an unverified DNNs-based controller.

Cite

Text

Zhong et al. "Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I12.26789

Markdown

[Zhong et al. "Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/zhong2023aaai-safe/) doi:10.1609/AAAI.V37I12.26789

BibTeX

@inproceedings{zhong2023aaai-safe,
  title     = {{Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games}},
  author    = {Zhong, Bingzhuo and Cao, Hongpeng and Zamani, Majid and Caccamo, Marco},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {15340-15349},
  doi       = {10.1609/AAAI.V37I12.26789},
  url       = {https://mlanthology.org/aaai/2023/zhong2023aaai-safe/}
}