Deep Learning the City: Quantifying Urban Perception at a Global Scale

Abstract

Computer vision methods that quantify the perception of urban environment are increasingly being used to study the relationship between a city’s physical appearance and the behavior and health of its residents. Yet, the throughput of current methods is too limited to quantify the perception of cities across the world. To tackle this challenge, we introduce a new crowdsourced dataset containing 110,988 images from 56 cities, and 1,170,000 pairwise comparisons provided by 81,630 online volunteers along six perceptual attributes: safe, lively, boring, wealthy, depressing, and beautiful. Using this data, we train a Siamese-like convolutional neural architecture, which learns from a joint classification and ranking loss, to predict human judgments of pairwise image comparisons. Our results show that crowdsourcing combined with neural networks can produce urban perception data at the global scale.

Cite

Text

Dubey et al. "Deep Learning the City: Quantifying Urban Perception at a Global Scale." European Conference on Computer Vision, 2016. doi:10.1007/978-3-319-46448-0_12

Markdown

[Dubey et al. "Deep Learning the City: Quantifying Urban Perception at a Global Scale." European Conference on Computer Vision, 2016.](https://mlanthology.org/eccv/2016/dubey2016eccv-deep/) doi:10.1007/978-3-319-46448-0_12

BibTeX

@inproceedings{dubey2016eccv-deep,
  title     = {{Deep Learning the City: Quantifying Urban Perception at a Global Scale}},
  author    = {Dubey, Abhimanyu and Naik, Nikhil and Parikh, Devi and Raskar, Ramesh and Hidalgo, César A.},
  booktitle = {European Conference on Computer Vision},
  year      = {2016},
  pages     = {196-212},
  doi       = {10.1007/978-3-319-46448-0_12},
  url       = {https://mlanthology.org/eccv/2016/dubey2016eccv-deep/}
}