Temporal Distance Matrices for Squat Classification

Abstract

When working out, it is necessary to perform the same action many times for it to have effect. If the action, such as squats or bench pressing, is performed with poor form, it can lead to serious injuries in the long term. For this purpose, we present an action dataset of squats where different types of poor form have been annotated with a diversity of users and backgrounds, and propose a model, based on temporal distance matrices, for the classification task. We first run a 3D pose detector, then we normalize the pose and compute the distance matrix, in which each element represents the distance between two joints. This representation is invariant to differences in individuals, global translation, and global rotation, allowing for high generalization to real world data. Our classification model consists of a CNN with 1D convolutions. Results show that our method significantly outperforms existing approaches for the task.

Cite

Text

Ogata et al. "Temporal Distance Matrices for Squat Classification." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019. doi:10.1109/CVPRW.2019.00309

Markdown

[Ogata et al. "Temporal Distance Matrices for Squat Classification." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/ogata2019cvprw-temporal/) doi:10.1109/CVPRW.2019.00309

BibTeX

@inproceedings{ogata2019cvprw-temporal,
  title     = {{Temporal Distance Matrices for Squat Classification}},
  author    = {Ogata, Ryoji and Simo-Serra, Edgar and Iizuka, Satoshi and Ishikawa, Hiroshi},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2019},
  pages     = {2533-2542},
  doi       = {10.1109/CVPRW.2019.00309},
  url       = {https://mlanthology.org/cvprw/2019/ogata2019cvprw-temporal/}
}