Exploiting Sparsity for Real Time Video Labelling

Abstract

Until recently, inference on fully connected graphs of pixel labels for scene understanding has been computationally expensive, so fast methods have focussed on neighbour connections and unary computation. However, with efficient CRF methods for inference on fully connected graphs, the opportunity exists for exploring other approaches. In this paper, we present a fast approach that calculates unary labels sparsely and relies on inference on fully connected graphs for label propagation. This reduces the unary computation which is now the most computationally expensive component. On a standard road scene dataset (CamVid), we show that accuarcy remains high when less than 0.15 percent of unary potentials are used. This achieves a reduction in computation by a factor of more than 750, with only small losses on global accuracy. This facilitates real-time processing on standard hardware that produces almost state-of-the-art results.

Cite

Text

Horne et al. "Exploiting Sparsity for Real Time Video Labelling." IEEE/CVF International Conference on Computer Vision Workshops, 2013. doi:10.1109/ICCVW.2013.87

Markdown

[Horne et al. "Exploiting Sparsity for Real Time Video Labelling." IEEE/CVF International Conference on Computer Vision Workshops, 2013.](https://mlanthology.org/iccvw/2013/horne2013iccvw-exploiting/) doi:10.1109/ICCVW.2013.87

BibTeX

@inproceedings{horne2013iccvw-exploiting,
  title     = {{Exploiting Sparsity for Real Time Video Labelling}},
  author    = {Horne, Lachlan and Alvarez, Jose M. and Barnes, Nick},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2013},
  pages     = {632-637},
  doi       = {10.1109/ICCVW.2013.87},
  url       = {https://mlanthology.org/iccvw/2013/horne2013iccvw-exploiting/}
}