ViSeR: Visual Self-Regularization
Abstract
We propose using large set of unlabeled images as a source of regularization data for learning robust representation. Given a visual model trained in a supervised fashion, we augment our training samples by incorporating large number of unlabeled data and train a semi-supervised model. We demonstrate that our proposed learning approach leverages an abundance of unlabeled images and boosts the visual recognition performance which alleviates the need to rely on large labeled datasets for learning robust representation. In our approach, each labeled image propagates its label to its nearest unlabeled image instances. These retrieved unlabeled images serve as local perturbations of each labeled image to perform Visual Self-Regularization (ViSeR). Using the labeled instances and our regularizers we show that we significantly improve object categorization and localization on the MS COCO and Visual Genome datasets.
Cite
Text
Izadinia and Garrigues. "ViSeR: Visual Self-Regularization." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020. doi:10.1109/CVPRW50498.2020.00479Markdown
[Izadinia and Garrigues. "ViSeR: Visual Self-Regularization." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020.](https://mlanthology.org/cvprw/2020/izadinia2020cvprw-viser/) doi:10.1109/CVPRW50498.2020.00479BibTeX
@inproceedings{izadinia2020cvprw-viser,
title = {{ViSeR: Visual Self-Regularization}},
author = {Izadinia, Hamid and Garrigues, Pierre},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2020},
pages = {4058-4063},
doi = {10.1109/CVPRW50498.2020.00479},
url = {https://mlanthology.org/cvprw/2020/izadinia2020cvprw-viser/}
}