Deep MANTA: A Coarse-to-Fine Many-Task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image
Abstract
In this paper, we present a novel approach, called Deep MANTA (Deep Many-Tasks), for many-task vehicle analysis from a given image. A robust convolutional network is introduced for simultaneous vehicle detection, part localization, visibility characterization and 3D dimension estimation. Its architecture is based on a new coarse-to-fine object proposal that boosts the vehicle detection. Moreover, the Deep MANTA network is able to localize vehicle parts even if these parts are not visible. In the inference, the network's outputs are used by a real time robust pose estimation algorithm for fine orientation estimation and 3D vehicle localization. We show in experiments that our method outperforms monocular state-of-the-art approaches on vehicle detection, orientation and 3D location tasks on the very challenging KITTI benchmark.
Cite
Text
Chabot et al. "Deep MANTA: A Coarse-to-Fine Many-Task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image." Conference on Computer Vision and Pattern Recognition, 2017. doi:10.1109/CVPR.2017.198Markdown
[Chabot et al. "Deep MANTA: A Coarse-to-Fine Many-Task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image." Conference on Computer Vision and Pattern Recognition, 2017.](https://mlanthology.org/cvpr/2017/chabot2017cvpr-deep/) doi:10.1109/CVPR.2017.198BibTeX
@inproceedings{chabot2017cvpr-deep,
title = {{Deep MANTA: A Coarse-to-Fine Many-Task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image}},
author = {Chabot, Florian and Chaouch, Mohamed and Rabarisoa, Jaonary and Teuliere, Celine and Chateau, Thierry},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2017},
doi = {10.1109/CVPR.2017.198},
url = {https://mlanthology.org/cvpr/2017/chabot2017cvpr-deep/}
}