Bootstrap AutoEncoders with Contrastive Paradigm for Self-Supervised Gaze Estimation
Abstract
Existing self-supervised methods for gaze estimation using the dominant streams of contrastive and generative approaches are restricted to eye images and could fail in general full-face settings. In this paper, we reveal that contrastive methods are ineffective in data augmentation for self-supervised full-face gaze estimation, while generative methods are prone to trivial solutions due to the absence of explicit regularization on semantic representations. To address this challenge, we propose a novel approach called Bootstrap auto-encoders with Contrastive paradigm (BeCa), which combines the strengths of both generative and contrastive methods. Specifically, we revisit the Auto-Encoder used in generative approaches and incorporate the contrastive paradigm to introduce explicit regularization on gaze representation. Furthermore, we design the InfoMSE loss as an alternative to the vanilla MSE loss for Auto-Encoder to mitigate the inconsistency between reconstruction and representation learning. Experimental results demonstrate that the proposed approaches outperform state-of-the-art unsupervised gaze approaches on extensive datasets (including wild scenes) under both within-dataset and cross-dataset protocols.
Cite
Text
Wang et al. "Bootstrap AutoEncoders with Contrastive Paradigm for Self-Supervised Gaze Estimation." International Conference on Machine Learning, 2024.Markdown
[Wang et al. "Bootstrap AutoEncoders with Contrastive Paradigm for Self-Supervised Gaze Estimation." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/wang2024icml-bootstrap/)BibTeX
@inproceedings{wang2024icml-bootstrap,
title = {{Bootstrap AutoEncoders with Contrastive Paradigm for Self-Supervised Gaze Estimation}},
author = {Wang, Yaoming and Li, Jin and Dai, Wenrui and Shi, Bowen and Zhang, Xiaopeng and Li, Chenglin and Xiong, Hongkai},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {50794-50806},
volume = {235},
url = {https://mlanthology.org/icml/2024/wang2024icml-bootstrap/}
}