Multi-View Feature Engineering and Learning

Abstract

We frame the problem of local representation of imaging data as the computation of minimal sufficient statistics that are invariant to nuisance variability induced by viewpoint and illumination. We show that, under very stringent conditions, these are related to "feature descriptors" commonly used in Computer Vision. Such conditions can be relaxed if multiple views of the same scene are available. We propose a sampling-based and a point-estimate based approximation of such a representation, compared empirically on image-to-(multiple)image matching, for which we introduce a multi-view wide-baseline matching benchmark, consisting of a mixture of real and synthetic objects with ground truth camera motion and dense three-dimensional geometry.

Cite

Text

Dong et al. "Multi-View Feature Engineering and Learning." Conference on Computer Vision and Pattern Recognition, 2015. doi:10.1109/CVPR.2015.7298945

Markdown

[Dong et al. "Multi-View Feature Engineering and Learning." Conference on Computer Vision and Pattern Recognition, 2015.](https://mlanthology.org/cvpr/2015/dong2015cvpr-multiview/) doi:10.1109/CVPR.2015.7298945

BibTeX

@inproceedings{dong2015cvpr-multiview,
  title     = {{Multi-View Feature Engineering and Learning}},
  author    = {Dong, Jingming and Karianakis, Nikolaos and Davis, Damek and Hernandez, Joshua and Balzer, Jonathan and Soatto, Stefano},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2015},
  doi       = {10.1109/CVPR.2015.7298945},
  url       = {https://mlanthology.org/cvpr/2015/dong2015cvpr-multiview/}
}