Enhancing Generalization of First-Order Meta-Learning

Abstract

In this study we focus on first-order meta-learning algorithms that aim to learn a parameter initialization of a network which can quickly adapt to new concepts, given a few examples. We investigate two approaches to enhance generalization and speed of learning of such algorithms, particularly expanding on the Reptile (Nichol et al., 2018) algorithm. We introduce a novel regularization technique called meta-step gradient pruning and also investigate the effects of increasing the depth of network architectures in first-order meta-learning. We present an empirical evaluation of both approaches, where we match benchmark few-shot image classification results with 10 times fewer iterations using Mini-ImageNet dataset and with the use of deeper networks, we attain accuracies that surpass the current benchmarks of few-shot image classification using Omniglot dataset.

Cite

Text

Jayathilaka. "Enhancing Generalization of First-Order Meta-Learning." ICLR 2019 Workshops: LLD, 2019.

Markdown

[Jayathilaka. "Enhancing Generalization of First-Order Meta-Learning." ICLR 2019 Workshops: LLD, 2019.](https://mlanthology.org/iclrw/2019/jayathilaka2019iclrw-enhancing/)

BibTeX

@inproceedings{jayathilaka2019iclrw-enhancing,
  title     = {{Enhancing Generalization of First-Order Meta-Learning}},
  author    = {Jayathilaka, Mirantha},
  booktitle = {ICLR 2019 Workshops: LLD},
  year      = {2019},
  url       = {https://mlanthology.org/iclrw/2019/jayathilaka2019iclrw-enhancing/}
}