Attentional Bottleneck: Towards an Interpretable Deep Driving Network
Abstract
Deep neural networks are a key component of behavior prediction and motion generation for self-driving cars. One of their main drawbacks is a lack of transparency: they should provide easy to interpret rationales for what triggers certain behaviors. We propose an architecture called Attentional Bottleneck with the goal of improving transparency. Our key idea is to combine visual attention, which identifies what aspects of the input the model is using, with an information bottleneck that enables the model to only use aspects of the input which are important. This not only provides sparse and interpretable attention maps (e.g. focusing only on specific vehicles in the scene), but it adds this transparency at no cost to model accuracy. In fact, we find slight improvements in accuracy when applying Attentional Bottleneck to the ChauffeurNet model in comparison to a traditional visual attention model that degrades accuracy.
Cite
Text
Kim and Bansal. "Attentional Bottleneck: Towards an Interpretable Deep Driving Network." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020. doi:10.1109/CVPRW50498.2020.00169Markdown
[Kim and Bansal. "Attentional Bottleneck: Towards an Interpretable Deep Driving Network." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020.](https://mlanthology.org/cvprw/2020/kim2020cvprw-attentional/) doi:10.1109/CVPRW50498.2020.00169BibTeX
@inproceedings{kim2020cvprw-attentional,
title = {{Attentional Bottleneck: Towards an Interpretable Deep Driving Network}},
author = {Kim, Jinkyu and Bansal, Mayank},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2020},
pages = {1310-1313},
doi = {10.1109/CVPRW50498.2020.00169},
url = {https://mlanthology.org/cvprw/2020/kim2020cvprw-attentional/}
}