Video-Based Localization Without 3D Mapping for the Visually Impaired

Abstract

In this paper, we present a system for indoor human localization that does not need 3D reconstruction of features or landmarks. We assume that a video sequence has been acquired and that keyframes have been registered with respect to 2D positions and orientations. In online mode, we use only a handheld monochrome fisheye camera and a synchronized IMU as sensory inputs. The query is not based on a single image but uses a HMM-based state estimator. Our image representation consists of initial global GIST vectors followed by local SURF features. We present a novel approach to localization by using search space reduction on global features, then HMM based position prediction and estimation on local features. Experimental results show that accurate localization is achieved and realtime performance is feasible. This work demonstrates that a working portable system could be designed for the visually impaired.

Cite

Text

Liu et al. "Video-Based Localization Without 3D Mapping for the Visually Impaired." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2010. doi:10.1109/CVPRW.2010.5543581

Markdown

[Liu et al. "Video-Based Localization Without 3D Mapping for the Visually Impaired." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2010.](https://mlanthology.org/cvprw/2010/liu2010cvprw-videobased/) doi:10.1109/CVPRW.2010.5543581

BibTeX

@inproceedings{liu2010cvprw-videobased,
  title     = {{Video-Based Localization Without 3D Mapping for the Visually Impaired}},
  author    = {Liu, Jason J. and Phillips, Cody J. and Daniilidis, Kostas},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2010},
  pages     = {23-30},
  doi       = {10.1109/CVPRW.2010.5543581},
  url       = {https://mlanthology.org/cvprw/2010/liu2010cvprw-videobased/}
}