model-agnostic meta-learning for fast adaptation of deep...

18
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks Chelsea Finn, Pieter Abbeel, Sergey Levine Presented by: Teymur Azayev CTU in Prague 17 January 2019

Upload: others

Post on 17-Aug-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Model-Agnostic Meta-Learning for FastAdaptation of Deep Networks

Chelsea Finn, Pieter Abbeel, Sergey LevinePresented by: Teymur Azayev

CTU in Prague

17 January 2019

Page 2: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Deep Learning

I Very powerful, expressive differentiable models.

I Flexibility is a double edged sword.

Page 3: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

How do we reduce the amount of required samples?

Use Use Prior knowledge (not in a Bayesian sense). This can be inthe form of:

I Model constraint

I Sampling strategy

I Update rule

I Loss function

I etc...

Page 4: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Meta learning

Learning to learn fast.Essentially learning a prior from a distribution of tasks.Several recent successful approaches:

I Model based meta-learning [Adam Santoro et al.],[Jx Wang et al.], [Yan Duan et al.]

I Metric meta-learning[Gregory Koch, Richard Zemel, and Ruslan Salakhutdinov.],[Oriol Vinyals et al.]

I Optimization based meta-learning[Sachin Ravi and Hugo Larochelle],[Marcin Andrychowicz et al.],

Page 5: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

MAML

Model Agnostic Metal Learning

Main idea: Learn a parameter initialization for a distribution oftasks, such that given a new task a small amount of examples(gradient updates) suffice.

Page 6: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Definitions

Task Ti ∼ p(T ) is defined as a tuple (Hi , qi ,LTi) consisting of

I time horizon Hi where for supervised learning Hi = 1

I initial state distribution qi (x0) and state transition distributionqi (xt+1|xt)

I Task loss function LTi→ R

I Task distribution p

Page 7: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Losses

I θ∗i is the optimal parameter for task Ti

I θ′i is the parameters obtained for task Ti after a single update

I 2) is the meta objective

Page 8: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Algorithm

Page 9: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Reinforcement learning

Page 10: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Reinforcement learning adaptation

Page 11: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Sin wave regression

Tasks: Regressing randomly generated sin waves

I amplitudes ranging in [0.1, 5]

I phases [0, 2π]

I Sampled uniformly in range [−5, 5]

Page 12: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Sin wave regression

Page 13: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Classification tasks

Omniglot

I 20 instances of 1623 characters from 50 different alphabets

I Each instance drawn by a different person

I Randomly select 1200 characters for training and theremaining for testing

MiniImagenet

I 64 training classes, 12 validation classes, and 24 test classes

Page 14: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

RL experiment

I Rllab benchmark suite, Mujoco simulator

I Gradient update are computed using policy gradientalgorithms.

I Tasks are defined by the agents simply having slightlydifferent goals

I Agents are expected to infer new goal from reward afterreceiving only 1 gradient update.

Page 15: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Conclusion

I Simple effective meta learning method

I Decent amount of follow up work [?], [?]

I Concept extendable to meta learning other parts of thetraining procedure

Page 16: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

Thank you for your attention

Page 17: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

ReferencesMarcin Andrychowicz et al.

Learning to learn by gradient descent by gradient descent.

NIPS 2016

Yan Duan et al.

RL2: Fast Reinforcement Learning via Slow Reinforcement Learning.

2016

Gregory Koch, Richard Zemel, and Ruslan Salakhutdinov.

Siamese Neural Networks for One-shot Image Recognition

ICML 2015

Zhenguo Li et al.

Meta-SGD: Learning to Learn quickly for few shot learning.

2017

Matthias Plappert et al.

Meta-SGD: Parameter Space Noise for Exploration

2017

Sachin Ravi and Hugo Larochelle

Meta-SGD:Optimization as a Model for Few-shot Learning

ICLR 2017

Adam Santoro et al.

One-shot Learning with Memory-Augmented Neural Networks

ICML 2016

Oriol Vinyals et al.

Matching Networks for One Shot Learning

NIPS 2016

Jx Wang et al.

Learning to Reinforcement Learn

2016

Risto Vuorio, Shao-Hua Sun, Hexiang Hu, Joseph J. Lim

Model-Agnostic Meta-Learning for Multimodal Task Distributions

2017

Ignasi Clavera, Anusha Nagabandi, Simin Liu, Ronald S. Fearing, PieterAbbeel, Sergey Levine, Chelsea Finn

Learning to Adapt in Dynamic, Real-World Environments throughMeta-Reinforcement Learning

2018

Page 18: Model-Agnostic Meta-Learning for Fast Adaptation of Deep ...cmp.felk.cvut.cz/~toliageo/rg/papers/slides/azayev_maml.pdf · One-shot Learning with Memory-Augmented Neural Networks

References I