+ All Categories
Home > Documents > via Task-Aware Modulation - GitHub Pages• Task network: further gradient adaptation via MAML steps...

via Task-Aware Modulation - GitHub Pages• Task network: further gradient adaptation via MAML steps...

Date post: 04-Jun-2020
Category:
Upload: others
View: 10 times
Download: 0 times
Share this document with a friend
1
Experiment - Regression Introduction Risto Vuorio* Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation Joseph J. Lim Shao-Hua Sun* Hexiang Hu Our Approach Experiment - Classification Experiment - Reinforcement Learning Experiment - Learned Task Embeddings Point Mass Reacher Ant Modulation Network Task Network x y ( ( K Samples Task Encoder υ Task Embedding Modulation Network Modulation Network MLPs x y 2 2 1 1 n n ˆ y Outer loop Task Encoder: produce the task embedding MLPs: modulate the task network blocks Inner loop Task network: fast adapt through gradient updates Parameters ! h ! g Intuition Modulation network: identify task modes and modulate the initialization accordingly Task network: further gradient adaptation via MAML steps Background Model-Agnostic Meta-Learning [1] Meta-learn a parameter initialization that can be fine-tuned for new tasks in few gradient update steps Inner loop Model-Agnostic Meta-Learning Objective Outer loop [1] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. "Model-agnostic meta-learning for fast adaptation of deep networks." in International Conference on Machine Learning 2017 Unimodal Task Distribution Multimodal Task Distribution Real-world task distributions are often multimodal Have a rich structure (e.g. multiple modes) Some knowledge can be transferable across modes/tasks Model-agnostic meta-learning (MAML) [1] Seek a common initialization parameter for all the modes An ensemble of MAMLs (Multi-MAML) Mode labels are often not available Prevent sharing related knowledge among modes/tasks
Transcript
Page 1: via Task-Aware Modulation - GitHub Pages• Task network: further gradient adaptation via MAML steps Background Model-Agnostic Meta-Learning [1] • Meta-learn a parameter initialization

Experiment - Regression

Introduction

Risto Vuorio*

Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

Joseph J. LimShao-Hua Sun* Hexiang Hu

Our Approach Experiment - Classification

Experiment - Reinforcement Learning

Experiment - Learned Task Embeddings

Poin

t Mas

sRe

ache

rAn

t

Modulation Network Task Network

x

y

( (

K⇥Samples

Task Encoder

�Task Embedding

Modulation Network

Modulation NetworkMLPs

x

y

�✓2⌧2

�✓1⌧1

�⌧n✓n

y

Outer loop• Task Encoder: produce the task embedding • MLPs: modulate the task network blocks

Inner loop• Task network: fast adapt through gradient updates

Parameters

!h

!g

Intuition• Modulation network: identify task modes and modulate the

initialization accordingly • Task network: further gradient adaptation via MAML steps

BackgroundModel-Agnostic Meta-Learning [1]

• Meta-learn a parameter initialization that can be fine-tuned for new tasks in few gradient update steps

• Inner loop

Model-Agnostic Meta-Learning Objective

• Outer loop

[1] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. "Model-agnostic meta-learning for fast adaptation of deep networks." in International Conference on Machine Learning 2017

Unimodal Task Distribution Multimodal Task Distribution

Real-world task distributions are often multimodal• Have a rich structure (e.g. multiple modes) • Some knowledge can be transferable across modes/tasks

Model-agnostic meta-learning (MAML) [1]• Seek a common initialization parameter for all the modes

An ensemble of MAMLs (Multi-MAML)• Mode labels are often not available • Prevent sharing related knowledge among modes/tasks

Recommended