On First-Order Meta-Learning Algorithms
Alex Nichol, Joshua Achiam, John Schulman
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/openai/supervised-reptileOfficialtf★ 0
- github.com/learnables/learn2learnpytorch★ 2,878
- github.com/sanowar-raihan/nerf-metapytorch★ 98
- github.com/Yuzhe-CHEN/NerfSNNpytorch★ 1
- github.com/aravindMahadevan/metaLearningAlgospytorch★ 0
- github.com/peisungtsai/Reptile-Pytorch-Implementationpytorch★ 0
- github.com/gebob19/cscd94_metalearningpytorch★ 0
- github.com/gabrielhuang/reptile-pytorchpytorch★ 0
- github.com/hfahrudin/reptile_implement_tf2tf★ 0
- github.com/MaximeVandegar/Papers-in-100-Lines-of-Code/tree/main/On_First_Order_Meta_Learning_Algorithmspytorch★ 0
Abstract
This paper considers meta-learning problems, where there is a distribution of tasks, and we would like to obtain an agent that performs well (i.e., learns quickly) when presented with a previously unseen task sampled from this distribution. We analyze a family of algorithms for learning a parameter initialization that can be fine-tuned quickly on a new task, using only first-order derivatives for the meta-learning updates. This family includes and generalizes first-order MAML, an approximation to MAML obtained by ignoring second-order derivatives. It also includes Reptile, a new algorithm that we introduce here, which works by repeatedly sampling a task, training on it, and moving the initialization towards the trained weights on that task. We expand on the results from Finn et al. showing that first-order meta-learning algorithms perform well on some well-established benchmarks for few-shot classification, and we provide theoretical analysis aimed at understanding why these algorithms work.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| Mini-Imagenet 10-way (1-shot) | Reptile+BN | Accuracy | 32 | — | Unverified |
| Mini-Imagenet 10-way (1-shot) | Reptile | Accuracy | 31.1 | — | Unverified |
| Mini-Imagenet 10-way (5-shot) | Reptile+BN | Accuracy | 47.6 | — | Unverified |
| Mini-Imagenet 10-way (5-shot) | Reptile | Accuracy | 44.7 | — | Unverified |
| Mini-Imagenet 5-way (1-shot) | Reptile + Transduction | Accuracy | 49.97 | — | Unverified |
| Mini-Imagenet 5-way (5-shot) | Reptile + Transduction | Accuracy | 65.99 | — | Unverified |
| OMNIGLOT - 1-Shot, 20-way | Reptile + Transduction | Accuracy | 89.43 | — | Unverified |
| OMNIGLOT - 1-Shot, 5-way | Reptile + Transduction | Accuracy | 97.68 | — | Unverified |
| OMNIGLOT - 5-Shot, 20-way | Reptile + Transduction | Accuracy | 97.12 | — | Unverified |
| OMNIGLOT - 5-Shot, 5-way | Reptile + Transduction | Accuracy | 99.48 | — | Unverified |
| Tiered ImageNet 10-way (1-shot) | Reptile | Accuracy | 33.7 | — | Unverified |
| Tiered ImageNet 10-way (1-shot) | Reptile+BN | Accuracy | 35.3 | — | Unverified |
| Tiered ImageNet 10-way (5-shot) | Reptile | Accuracy | 48 | — | Unverified |
| Tiered ImageNet 10-way (5-shot) | Reptile+BN | Accuracy | 52 | — | Unverified |