See Your Emotion from Gait Using Unlabeled Skeleton Data
Abstract
This paper focuses on contrastive learning for gait-based emotion recognition. The existing contrastive learning approaches are rarely suitable for learning skeleton-based gait representations, which suffer from limited gait diversity and inconsistent semantics. In this paper, we propose a Cross-coordinate contrastive learning framework utilizing Ambiguity samples for self-supervised Gait-based Emotion representation (CAGE). First, we propose ambiguity transform to push positive samples into ambiguous semantic space. By learning similarities between ambiguity samples and positive samples, our model can learn higher-level semantics of the gait sequences and maintain semantic diversity. Second, to encourage learning the semantic invariance, we uniquely propose cross-coordinate contrastive learning between the Cartesian coordinate and the Spherical coordinate, which brings rich supervisory signals to learn the intrinsic semantic consistency information. Exhaustive experiments show that CAGE improves existing self-supervised methods by 5%–10% accuracy, and it achieves comparable or even superior performance to supervised methods.
Cite
Text
Lu et al. "See Your Emotion from Gait Using Unlabeled Skeleton Data." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I2.25272Markdown
[Lu et al. "See Your Emotion from Gait Using Unlabeled Skeleton Data." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/lu2023aaai-see/) doi:10.1609/AAAI.V37I2.25272BibTeX
@inproceedings{lu2023aaai-see,
title = {{See Your Emotion from Gait Using Unlabeled Skeleton Data}},
author = {Lu, Haifeng and Hu, Xiping and Hu, Bin},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {1826-1834},
doi = {10.1609/AAAI.V37I2.25272},
url = {https://mlanthology.org/aaai/2023/lu2023aaai-see/}
}