Neural Inertial Localization

Abstract

This paper proposes the inertial localization problem, the task of estimating the absolute location from a sequence of inertial sensor measurements. This is an exciting and unexplored area of indoor localization research, where we present a rich dataset with 53 hours of inertial sensor data and the associated ground truth locations. We developed a solution, dubbed neural inertial localization (NILoc) which 1) uses a neural inertial navigation technique to turn inertial sensor history to a sequence of velocity vectors; then 2) employs a transformer-based neural architecture to find the device location from the sequence of velocity estimates. We only use an IMU sensor, which is energy efficient and privacy-preserving compared to WiFi, cameras, and other data sources. Our approach is significantly faster and achieves competitive results even compared with state-of-the-art methods that require a floorplan and run 20 to 30 times slower. We share our code, model and data at https://sachini.github.io/niloc.

Cite

Text

Herath et al. "Neural Inertial Localization." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00649

Markdown

[Herath et al. "Neural Inertial Localization." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/herath2022cvpr-neural/) doi:10.1109/CVPR52688.2022.00649

BibTeX

@inproceedings{herath2022cvpr-neural,
  title     = {{Neural Inertial Localization}},
  author    = {Herath, Sachini and Caruso, David and Liu, Chen and Chen, Yufan and Furukawa, Yasutaka},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {6604-6613},
  doi       = {10.1109/CVPR52688.2022.00649},
  url       = {https://mlanthology.org/cvpr/2022/herath2022cvpr-neural/}
}