Inertial-Vision: Cross-Domain Knowledge Transfer for Wearable Sensors

Abstract

Multi-modal ego-centric data from inertial measurement units (IMU) and first-person videos (FPV) can be effectively fused to recognise proprioceptive activities. Existing IMU-based approaches mostly employ cascades of handcrafted triaxial motion features or deep frameworks trained on limited data. FPV approaches generally encode scene dynamics with motion and pooled appearance features. In this paper, we propose a multi-modal ego-centric proprioceptive activity recognition that uses a convolutional neural network (CNN) followed by a long short-term memory (LSTM) network, transfer learning and a merit-based fusion of IMU and/or FPV streams. The CNN encodes short-term temporal dynamics of the ego-motion and the LSTM exploits the long-term temporal dependency among activities. The merit of a stream is evaluated with a sparsity measure of its initial classification output. We validate the proposed framework on multiple visual and inertial datasets.

Cite

Text

Abebe and Cavallaro. "Inertial-Vision: Cross-Domain Knowledge Transfer for Wearable Sensors." IEEE/CVF International Conference on Computer Vision Workshops, 2017. doi:10.1109/ICCVW.2017.165

Markdown

[Abebe and Cavallaro. "Inertial-Vision: Cross-Domain Knowledge Transfer for Wearable Sensors." IEEE/CVF International Conference on Computer Vision Workshops, 2017.](https://mlanthology.org/iccvw/2017/abebe2017iccvw-inertialvision/) doi:10.1109/ICCVW.2017.165

BibTeX

@inproceedings{abebe2017iccvw-inertialvision,
  title     = {{Inertial-Vision: Cross-Domain Knowledge Transfer for Wearable Sensors}},
  author    = {Abebe, Girmaw and Cavallaro, Andrea},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2017},
  pages     = {1392-1400},
  doi       = {10.1109/ICCVW.2017.165},
  url       = {https://mlanthology.org/iccvw/2017/abebe2017iccvw-inertialvision/}
}