SOTAVerified

Meta-SGD: Learning to Learn Quickly for Few-Shot Learning

2017-07-31Code Available1· sign in to hype

Zhenguo Li, Fengwei Zhou, Fei Chen, Hang Li

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Few-shot learning is challenging for learning algorithms that learn each task in isolation and from scratch. In contrast, meta-learning learns from many related tasks a meta-learner that can learn a new task more accurately and faster with fewer examples, where the choice of meta-learners is crucial. In this paper, we develop Meta-SGD, an SGD-like, easily trainable meta-learner that can initialize and adapt any differentiable learner in just one step, on both supervised learning and reinforcement learning. Compared to the popular meta-learner LSTM, Meta-SGD is conceptually simpler, easier to implement, and can be learned more efficiently. Compared to the latest meta-learner MAML, Meta-SGD has a much higher capacity by learning to learn not just the learner initialization, but also the learner update direction and learning rate, all in a single meta-learning process. Meta-SGD shows highly competitive performance for few-shot learning on regression, classification, and reinforcement learning.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Mini-Imagenet 20-way (1-shot)Meta SGDAccuracy17.56Unverified
Mini-Imagenet 20-way (1-shot)Matching Nets, (from )Accuracy17.31Unverified
Mini-Imagenet 20-way (1-shot)Meta LSTM, (from )Accuracy16.7Unverified
Mini-Imagenet 20-way (1-shot)MAML, (from )Accuracy16.49Unverified
Mini-Imagenet 20-way (5-shot)Meta SGDAccuracy28.92Unverified
Mini-Imagenet 20-way (5-shot)Meta LSTM, (from )Accuracy26.06Unverified
Mini-Imagenet 20-way (5-shot)Matching Nets, (from )Accuracy22.69Unverified
Mini-Imagenet 20-way (5-shot)MAML, (from )Accuracy19.29Unverified

Reproductions