Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
By Chelsea Finn, Pieter Abbeel, and Sergey Levine, ICML 2017
Present by: Shihao Ma, Yichun Zhang, and Zilun Zhang
CSC 2541 Winter 2021
April 1st, 2021
Agenda
Background: Meta Learning
Background: Meta Learning
Background: Few-Shot Learning
Background: Few-Shot Learning
MAML: Generic Definition
Learning Task (generic notation across different learning problems):
task specific loss
distribution over initial observations
transition distribution
episode length ( In i.i.d. supervised learning problems, the length )
output of model at time t
MAML: Algorithm Explanation
Inner Loop:
Outer Loop:
The MAML meta-gradient update involves a gradient through a gradient, which requires to compute Hessian vector product
Case Study 1: Classification
Data
Objective
Backbone - Conv4
Loss
Case Study 1: Classification
The model learned with MAML uses fewer overall parameters compared to sota models (in 2017), since the algorithm does not introduce additional parameters beyond the weights of the classifier itself
Split (train-val-test)
Case Study 2: Sine Wave Regression
Data
Objective
Model
Loss
Evaluation Protocol
Case Study 2: Sine Wave Regression
Case Study 3: Reinforcement Learning
Policy Gradient for discrete data (non-differentiable reward)
Related Works
Meta-Learning Application
Related Works
Few-shot Learning
Related Works
Differences with MAML
Problems of MAML
Previous: Update outer loop when completed all inner-loop updates
New: Update outer loop after every step of the inner-loop
Annealed weighting for the pre step losses
Use first-order approximated gradients for the first 50 epochs, then switch to second-order gradients for the remainder of the training phase.
Learn a learning rate and direction for each layer in the network as well as learn different learning rates for each adaptation of the base-network as it takes steps.
Learn a set of biases per-step within the inner-loop update process.
Increase convergence speed, stability and generalization performance.
Pre-training + Fine-Tuning V.S. MAML
MAML
Pre-training + Fine-Tuning
Summary