Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games
Abstract
Nowadays, AI-based techniques, such as deep neural networks (DNNs), are widely deployed in autonomous systems for complex mission requirements (e.g., motion planning in robotics). However, DNNs-based controllers are typically very complex, and it is very hard to formally verify their correctness, potentially causing severe risks for safety-critical autonomous systems. In this paper, we propose a construction scheme for a so-called Safe-visor architecture to sandbox DNNs-based controllers. Particularly, we consider the construction under a stochastic game framework to provide a system-level safety guarantee which is robust to noises and disturbances. A supervisor is built to check the control inputs provided by a DNNs-based controller and decide whether to accept them. Meanwhile, a safety advisor is running in parallel to provide fallback control inputs in case the DNN-based controller is rejected. We demonstrate the proposed approaches on a quadrotor employing an unverified DNNs-based controller.
Cite
Text
Zhong et al. "Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I12.26789Markdown
[Zhong et al. "Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/zhong2023aaai-safe/) doi:10.1609/AAAI.V37I12.26789BibTeX
@inproceedings{zhong2023aaai-safe,
title = {{Towards Safe AI: Sandboxing DNNs-Based Controllers in Stochastic Games}},
author = {Zhong, Bingzhuo and Cao, Hongpeng and Zamani, Majid and Caccamo, Marco},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {15340-15349},
doi = {10.1609/AAAI.V37I12.26789},
url = {https://mlanthology.org/aaai/2023/zhong2023aaai-safe/}
}