Tri-Net for Semi-Supervised Deep Learning
Abstract
Deep neural networks have witnessed great successes in various real applications, but it requires a large number of labeled data for training. In this paper, we propose tri-net, a deep neural network which is able to use massive unlabeled data to help learning with limited labeled data. We consider model initialization, diversity augmentation and pseudo-label editing simultaneously. In our work, we utilize output smearing to initialize modules, use fine-tuning on labeled data to augment diversity and eliminate unstable pseudo-labels to alleviate the influence of suspicious pseudo-labeled data. Experiments show that our method achieves the best performance in comparison with state-of-the-art semi-supervised deep learning methods. In particular, it achieves 8.30% error rate on CIFAR-10 by using only 4000 labeled examples.
Cite
Text
Chen et al. "Tri-Net for Semi-Supervised Deep Learning." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/278Markdown
[Chen et al. "Tri-Net for Semi-Supervised Deep Learning." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/chen2018ijcai-tri/) doi:10.24963/IJCAI.2018/278BibTeX
@inproceedings{chen2018ijcai-tri,
title = {{Tri-Net for Semi-Supervised Deep Learning}},
author = {Chen, Dongdong and Wang, Wei and Gao, Wei and Zhou, Zhi-Hua},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2018},
pages = {2014-2020},
doi = {10.24963/IJCAI.2018/278},
url = {https://mlanthology.org/ijcai/2018/chen2018ijcai-tri/}
}