KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder

Abstract

In this work we attempted to extend the thought and showcase a way forward for the Self-supervised Learning (SSL) learning paradigm by combining contrastive learning self-distillation (knowledge distillation) and masked data modelling the three major SSL frameworks to learn a joint and coordinated representation. The proposed technique of SSL learns by the collaborative power of different learning objectives of SSL. Hence to jointly learn the different SSL objectives we proposed a new SSL architecture KDC-MAE a complementary masking strategy to learn the modular correspondence and a weighted way to combine them coordinately. Experimental results conclude that the contrastive masking correspondence along with the KD learning objective has lent a hand to performing better learning for multiple modalities over multiple tasks.

Cite

Text

Bora et al. "KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder." Winter Conference on Applications of Computer Vision, 2025.

Markdown

[Bora et al. "KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder." Winter Conference on Applications of Computer Vision, 2025.](https://mlanthology.org/wacv/2025/bora2025wacv-kdcmae/)

BibTeX

@inproceedings{bora2025wacv-kdcmae,
  title     = {{KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder}},
  author    = {Bora, Maheswar and Atreya, Saurabh and Mukherjee, Aritra and Das, Abhijit},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2025},
  pages     = {7511-7521},
  url       = {https://mlanthology.org/wacv/2025/bora2025wacv-kdcmae/}
}