Proximate Sensing: Inferring What-Is-Where from Georeferenced Photo Collections

Abstract

The primary and novel contribution of this work is the conjecture that large collections of georeferenced photo collections can be used to derive maps of what-is-where on the surface of the earth. We investigate the application of what we term "proximate sensing" to the problem of land cover classification for a large geographic region. We show that our approach is able to achieve almost 75% classification accuracy in a binary land cover labelling problem using images from a photo sharing site in a completely automated fashion. We also investigate 1) how existing geographic knowledge can be used to provide labelled training data in a weakly-supervised manner; 2) the effect of the photographer's intent when he or she captures the photograph; and 3) a method for filtering out non-informative images.

Cite

Text

Leung and Newsam. "Proximate Sensing: Inferring What-Is-Where from Georeferenced Photo Collections." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2010. doi:10.1109/CVPR.2010.5540040

Markdown

[Leung and Newsam. "Proximate Sensing: Inferring What-Is-Where from Georeferenced Photo Collections." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2010.](https://mlanthology.org/cvpr/2010/leung2010cvpr-proximate/) doi:10.1109/CVPR.2010.5540040

BibTeX

@inproceedings{leung2010cvpr-proximate,
  title     = {{Proximate Sensing: Inferring What-Is-Where from Georeferenced Photo Collections}},
  author    = {Leung, Daniel and Newsam, Shawn D.},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2010},
  pages     = {2955-2962},
  doi       = {10.1109/CVPR.2010.5540040},
  url       = {https://mlanthology.org/cvpr/2010/leung2010cvpr-proximate/}
}