DeepNav: Learning to Navigate Large Cities

Abstract

We present DeepNav, a Convolutional Neural Network (CNN) based algorithm for navigating large cities using locally visible street-view images. The DeepNav agent learns to reach its destination quickly by making the correct navigation decisions at intersections. We collect a large-scale dataset of street-view images organized in a graph where nodes are connected by roads. This dataset contains 10 city graphs and a total of more than 1 million street-view images. We propose 3 supervised learning approaches for the navigation task, and show how A* search in the city graph can be used to generate labels for the images. Our annotation process is fully automated using publicly available mapping services, and requires no human input. We evaluate the proposed DeepNav models on 4 held-out cities for navigating to 5 different types of destinations and show that our algorithms outperform previous work that uses hand-crafted features and Support Vector Regression (SVR).

Cite

Text

Brahmbhatt and Hays. "DeepNav: Learning to Navigate Large Cities." Conference on Computer Vision and Pattern Recognition, 2017. doi:10.1109/CVPR.2017.329

Markdown

[Brahmbhatt and Hays. "DeepNav: Learning to Navigate Large Cities." Conference on Computer Vision and Pattern Recognition, 2017.](https://mlanthology.org/cvpr/2017/brahmbhatt2017cvpr-deepnav/) doi:10.1109/CVPR.2017.329

BibTeX

@inproceedings{brahmbhatt2017cvpr-deepnav,
  title     = {{DeepNav: Learning to Navigate Large Cities}},
  author    = {Brahmbhatt, Samarth and Hays, James},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2017},
  doi       = {10.1109/CVPR.2017.329},
  url       = {https://mlanthology.org/cvpr/2017/brahmbhatt2017cvpr-deepnav/}
}