A Group-Theoretic Framework for Data Augmentation
Abstract
Data augmentation has become an important part of modern deep learning pipelines and is typically needed to achieve state of the art performance for many learning tasks. It utilizes invariant transformations of the data, such as rotation, scale, and color shift, and the transformed images are added to the training set. However, these transformations are often chosen heuristically and a clear theoretical framework to explain the performance benefits of data augmentation is not available. In this paper, we develop such a framework to explain data augmentation as averaging over the orbits of the group that keeps the data distribution approximately invariant, and show that it leads to variance reduction. We study finite-sample and asymptotic empirical risk minimization and work out as examples the variance reduction in certain two-layer neural networks. We further propose a strategy to exploit the benefits of data augmentation for general learning tasks.
Cite
Text
Chen et al. "A Group-Theoretic Framework for Data Augmentation." Neural Information Processing Systems, 2020.Markdown
[Chen et al. "A Group-Theoretic Framework for Data Augmentation." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/chen2020neurips-grouptheoretic/)BibTeX
@inproceedings{chen2020neurips-grouptheoretic,
title = {{A Group-Theoretic Framework for Data Augmentation}},
author = {Chen, Shuxiao and Dobriban, Edgar and Lee, Jane},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/chen2020neurips-grouptheoretic/}
}