PRIMUS: Pretraining IMU Encoders with Multimodal Self-Supervision

Abstract

Sensing human motions through Inertial Measurement Units (IMUs) embedded in personal devices has enabled significant applications in health and wellness. While labeled IMU data is scarce, we can collect unlabeled or weakly labeled IMU data to model human motions. For video or text modalities, the "pretrain and adapt" approach utilizes large volumes of unlabeled or weakly labeled data for pretraining, building a strong feature extractor, followed by adaptation to specific tasks using limited labeled data. However, for IMU data, pretraining methods are poorly understood, and pretraining pipelines are rarely evaluated on out-of-domain tasks. We propose PRIMUS: a method for PRetraining IMU encoderS that uses a novel pretraining objective that is empirically validated based on downstream performance on both in-domain and out-of-domain datasets. The PRIMUS objective effectively enhances downstream performance by combining self-supervision, multimodal, and nearest-neighbor supervision. With fewer than 500 labeled samples per class, PRIMUS can improve test accuracy by up to 15\%, compared to state-of-the-art baselines. To benefit the broader community, we open-source our code at github.com/nokia-bell-labs/pretrained-imu-encoders.

Cite

Text

Das et al. "PRIMUS: Pretraining IMU Encoders with Multimodal Self-Supervision." NeurIPS 2024 Workshops: TSALM, 2024.

Markdown

[Das et al. "PRIMUS: Pretraining IMU Encoders with Multimodal Self-Supervision." NeurIPS 2024 Workshops: TSALM, 2024.](https://mlanthology.org/neuripsw/2024/das2024neuripsw-primus/)

BibTeX

@inproceedings{das2024neuripsw-primus,
  title     = {{PRIMUS: Pretraining IMU Encoders with Multimodal Self-Supervision}},
  author    = {Das, Arnav Mohanty and Tang, Chi Ian and Kawsar, Fahim and Malekzadeh, Mohammad},
  booktitle = {NeurIPS 2024 Workshops: TSALM},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/das2024neuripsw-primus/}
}