Joint Prediction of Amodal and Visible Semantic Segmentation for Automated Driving
Abstract
Amodal perception is the ability to hallucinate full shapes of (partially) occluded objects. While natural to humans, learning-based perception methods often only focus on the visible parts of scenes. This constraint is critical for safe automated driving since detection capabilities of perception methods are limited when faced with (partial) occlusions. Moreover, corner cases can emerge from occlusions while the perception method is oblivious. In this work, we investigate the possibilities of joint prediction of amodal and visible semantic segmentation masks. More precisely, we investigate whether both perception tasks benefit from a joint training approach. We report our findings on both the Cityscapes and the Amodal Cityscapes dataset. The proposed joint training outperforms the separately trained networks in terms of mean intersection over union in amodal areas of the masks by $6.84\%$ 6.84 % absolute, while even slightly improving the visible segmentation performance.
Cite
Text
Breitenstein et al. "Joint Prediction of Amodal and Visible Semantic Segmentation for Automated Driving." European Conference on Computer Vision Workshops, 2022. doi:10.1007/978-3-031-25056-9_40Markdown
[Breitenstein et al. "Joint Prediction of Amodal and Visible Semantic Segmentation for Automated Driving." European Conference on Computer Vision Workshops, 2022.](https://mlanthology.org/eccvw/2022/breitenstein2022eccvw-joint/) doi:10.1007/978-3-031-25056-9_40BibTeX
@inproceedings{breitenstein2022eccvw-joint,
title = {{Joint Prediction of Amodal and Visible Semantic Segmentation for Automated Driving}},
author = {Breitenstein, Jasmin and Löhdefink, Jonas and Fingscheidt, Tim},
booktitle = {European Conference on Computer Vision Workshops},
year = {2022},
pages = {633-645},
doi = {10.1007/978-3-031-25056-9_40},
url = {https://mlanthology.org/eccvw/2022/breitenstein2022eccvw-joint/}
}