Efficient Semantic Video Segmentation with Per-Frame Inference
Abstract
For semantic segmentation, most existing real-time deep mod-els trained with each frame independently may produce inconsistent results when tested on a video sequence. A few methods take the correlations in the video sequence into account, e.g., by propagating the results to the neighboring frames using optical flow or extracting frame representations using multi-frame information, which may lead to inaccurate results or unbalanced latency. In contrast, here we explicitly consider the temporal consistency among frames as extra constraints during training and process each frame independently in the inference phase. Thus no computation overhead is introduced for inference. Compact models are employed for real-time execution. To narrow the performance gap between compact models and large models, new temporal knowledge distillation methods are designed. Weighing among accuracy, temporal smoothness, and efficiency, our proposed method outperforms previous keyframe based methods and corresponding baselines which are trained with each frame independently on benchmark datasets including cityscapes and Camvid.
Cite
Text
Liu et al. "Efficient Semantic Video Segmentation with Per-Frame Inference." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58607-2_21Markdown
[Liu et al. "Efficient Semantic Video Segmentation with Per-Frame Inference." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/liu2020eccv-efficient/) doi:10.1007/978-3-030-58607-2_21BibTeX
@inproceedings{liu2020eccv-efficient,
title = {{Efficient Semantic Video Segmentation with Per-Frame Inference}},
author = {Liu, Yifan and Shen, Chunhua and Yu, Changqian and Wang, Jingdong},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58607-2_21},
url = {https://mlanthology.org/eccv/2020/liu2020eccv-efficient/}
}