KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder
Abstract
In this work we attempted to extend the thought and showcase a way forward for the Self-supervised Learning (SSL) learning paradigm by combining contrastive learning self-distillation (knowledge distillation) and masked data modelling the three major SSL frameworks to learn a joint and coordinated representation. The proposed technique of SSL learns by the collaborative power of different learning objectives of SSL. Hence to jointly learn the different SSL objectives we proposed a new SSL architecture KDC-MAE a complementary masking strategy to learn the modular correspondence and a weighted way to combine them coordinately. Experimental results conclude that the contrastive masking correspondence along with the KD learning objective has lent a hand to performing better learning for multiple modalities over multiple tasks.
Cite
Text
Bora et al. "KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder." Winter Conference on Applications of Computer Vision, 2025.Markdown
[Bora et al. "KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder." Winter Conference on Applications of Computer Vision, 2025.](https://mlanthology.org/wacv/2025/bora2025wacv-kdcmae/)BibTeX
@inproceedings{bora2025wacv-kdcmae,
title = {{KDC-MAE: Knowledge Distilled Contrastive Mask Auto-Encoder}},
author = {Bora, Maheswar and Atreya, Saurabh and Mukherjee, Aritra and Das, Abhijit},
booktitle = {Winter Conference on Applications of Computer Vision},
year = {2025},
pages = {7511-7521},
url = {https://mlanthology.org/wacv/2025/bora2025wacv-kdcmae/}
}