Tri-Net for Semi-Supervised Deep Learning

Abstract

Deep neural networks have witnessed great successes in various real applications, but it requires a large number of labeled data for training. In this paper, we propose tri-net, a deep neural network which is able to use massive unlabeled data to help learning with limited labeled data. We consider model initialization, diversity augmentation and pseudo-label editing simultaneously. In our work, we utilize output smearing to initialize modules, use fine-tuning on labeled data to augment diversity and eliminate unstable pseudo-labels to alleviate the influence of suspicious pseudo-labeled data. Experiments show that our method achieves the best performance in comparison with state-of-the-art semi-supervised deep learning methods. In particular, it achieves 8.30% error rate on CIFAR-10 by using only 4000 labeled examples.

Cite

Text

Chen et al. "Tri-Net for Semi-Supervised Deep Learning." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/278

Markdown

[Chen et al. "Tri-Net for Semi-Supervised Deep Learning." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/chen2018ijcai-tri/) doi:10.24963/IJCAI.2018/278

BibTeX

@inproceedings{chen2018ijcai-tri,
  title     = {{Tri-Net for Semi-Supervised Deep Learning}},
  author    = {Chen, Dongdong and Wang, Wei and Gao, Wei and Zhou, Zhi-Hua},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {2014-2020},
  doi       = {10.24963/IJCAI.2018/278},
  url       = {https://mlanthology.org/ijcai/2018/chen2018ijcai-tri/}
}