Slimmed Asymmetrical Contrastive Learning and Cross Distillation for Lightweight Model Training
Abstract
Contrastive learning (CL) has been widely investigated with various learning mechanisms and achieves strong capability in learning representations of data in a self-supervised manner using unlabeled data. A common fashion of contrastive learning on this line is employing mega-sized encoders to achieve comparable performance as the supervised learning counterpart. Despite the success of the labelless training, current contrastive learning algorithms *failed* to achieve good performance with lightweight (compact) models, e.g., MobileNet, while the requirements of the heavy encoders impede the energy-efficient computation, especially for resource-constrained AI applications. Motivated by this, we propose a new self-supervised CL scheme, named SACL-XD, consisting of two technical components, **S**limmed **A**symmetrical **C**ontrastive **L**earning (SACL) and **Cross**-**D**istillation (XD), which collectively enable efficient CL with compact models. While relevant prior works employed a strong pre-trained model as the teacher of unsupervised knowledge distillation to a lightweight encoder, our proposed method trains CL models from scratch and outperforms them even without such an expensive requirement. Compared to the SoTA lightweight CL training (distillation) algorithms, SACL-XD achieves 1.79% ImageNet-1K accuracy improvement on MobileNet-V3 with 64$\times$ training FLOPs reduction.
Cite
Text
Meng et al. "Slimmed Asymmetrical Contrastive Learning and Cross Distillation for Lightweight Model Training." Neural Information Processing Systems, 2023.Markdown
[Meng et al. "Slimmed Asymmetrical Contrastive Learning and Cross Distillation for Lightweight Model Training." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/meng2023neurips-slimmed/)BibTeX
@inproceedings{meng2023neurips-slimmed,
title = {{Slimmed Asymmetrical Contrastive Learning and Cross Distillation for Lightweight Model Training}},
author = {Meng, Jian and Yang, Li and Lee, Kyungmin and Shin, Jinwoo and Fan, Deliang and Seo, Jae-sun},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/meng2023neurips-slimmed/}
}