AAR-CNNs: Auto Adaptive Regularized Convolutional Neural Networks

Abstract

In order to address the overfitting problem caused by the small or simple training datasets and the large model’s size in Convolutional Neural Networks (CNNs), a novel Auto Adaptive Regularization (AAR) method is proposed in this paper. The relevant networks can be called AAR-CNNs. AAR is the first method using the “abstraction extent” (predicted by AE net) and a tiny learnable module (SE net) to auto adaptively predict more accurate and individualized regularization information. The AAR module can be directly inserted into every stage of any popular networks and trained end to end to improve the networks’ flexibility. This method can not only regularize the network at both the forward and the backward processes in the training phase, but also regularize the network on a more refined level (channel or pixel level) depending on the abstraction extent’s form. Comparative experiments are performed on low resolution ImageNet, CIFAR and SVHN datasets. Experimental results show that the AAR-CNNs can achieve state-of-the-art performances on these datasets.

Cite

Text

Lu et al. "AAR-CNNs: Auto Adaptive Regularized Convolutional Neural Networks." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/348

Markdown

[Lu et al. "AAR-CNNs: Auto Adaptive Regularized Convolutional Neural Networks." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/lu2018ijcai-aar/) doi:10.24963/IJCAI.2018/348

BibTeX

@inproceedings{lu2018ijcai-aar,
  title     = {{AAR-CNNs: Auto Adaptive Regularized Convolutional Neural Networks}},
  author    = {Lu, Yao and Lu, Guangming and Xu, Yuanrong and Zhang, Bob},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {2511-2517},
  doi       = {10.24963/IJCAI.2018/348},
  url       = {https://mlanthology.org/ijcai/2018/lu2018ijcai-aar/}
}