DAG-Recurrent Neural Networks for Scene Labeling
Abstract
In image labeling, local representations for image units (pixels, patches or superpixels) are usually generated from their surrounding image patches, thus long-range contextual information is not effectively encoded. In this paper, we introduce recurrent neural networks (RNNs) to address this issue. Specifically, directed acyclic graph RNNs (DAG-RNNs) are proposed to process DAG-structured images, which enables the network to model long-range semantic dependencies among image units. Our DAG-RNNs are capable of tremendously enhancing the discriminative power of local representations, which significantly benefits the local classification. Meanwhile, we propose a novel class weighting function that attends to rare classes, which phenomenally boosts the recognition accuracy for non-frequent classes. Integrating with convolution and deconvolution layers, our DAG-RNNs achieve new state-of-the-art results on the challenging SiftFlow, CamVid and Barcelona benchmarks.
Cite
Text
Shuai et al. "DAG-Recurrent Neural Networks for Scene Labeling." Conference on Computer Vision and Pattern Recognition, 2016. doi:10.1109/CVPR.2016.394Markdown
[Shuai et al. "DAG-Recurrent Neural Networks for Scene Labeling." Conference on Computer Vision and Pattern Recognition, 2016.](https://mlanthology.org/cvpr/2016/shuai2016cvpr-dagrecurrent/) doi:10.1109/CVPR.2016.394BibTeX
@inproceedings{shuai2016cvpr-dagrecurrent,
title = {{DAG-Recurrent Neural Networks for Scene Labeling}},
author = {Shuai, Bing and Zuo, Zhen and Wang, Bing and Wang, Gang},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2016},
doi = {10.1109/CVPR.2016.394},
url = {https://mlanthology.org/cvpr/2016/shuai2016cvpr-dagrecurrent/}
}