Adversarial Learning of General Transformations for Data Augmentation
Abstract
Data augmentation (DA) is fundamental against overfitting in large convolutional neural networks, especially with a limited training dataset. In images, DA is usually based on heuristic transformations, like geometric or color transformations. Instead of using predefined transformations, our work learns data augmentation directly from the training data by learning to transform images with an encoder-decoder architecture combined with a spatial transformer network. The transformed images still belong to the same class, but are new, more complex samples for the classifier. Our experiments show that our approach is better than previous generative data augmentation methods, and comparable to predefined transformation methods when training an image classifier.
Cite
Text
Mounsaveng et al. "Adversarial Learning of General Transformations for Data Augmentation." ICLR 2019 Workshops: LLD, 2019.Markdown
[Mounsaveng et al. "Adversarial Learning of General Transformations for Data Augmentation." ICLR 2019 Workshops: LLD, 2019.](https://mlanthology.org/iclrw/2019/mounsaveng2019iclrw-adversarial/)BibTeX
@inproceedings{mounsaveng2019iclrw-adversarial,
title = {{Adversarial Learning of General Transformations for Data Augmentation}},
author = {Mounsaveng, Saypraseuth and Vazquez, David and Ayed, Ismail Ben and Pedersoli, Marco},
booktitle = {ICLR 2019 Workshops: LLD},
year = {2019},
url = {https://mlanthology.org/iclrw/2019/mounsaveng2019iclrw-adversarial/}
}