Learning to Generalize to Unseen Tasks with Bilevel Optimization

08/05/2019
by   Hayeon Lee, et al.
0

Recent metric-based meta-learning approaches, which learn a metric space that generalizes well over combinatorial number of different classification tasks sampled from a task distribution, have been shown to be effective for few-shot classification tasks of unseen classes. They are often trained with episodic training where they iteratively train a common metric space that reduces distance between the class representatives and instances belonging to each class, over large number of episodes with random classes. However, this training is limited in that while the main target is the generalization to the classification of unseen classes during training, there is no explicit consideration of generalization during meta-training phase. To tackle this issue, we propose a simple yet effective meta-learning framework for metricbased approaches, which we refer to as learning to generalize (L2G), that explicitly constrains the learning on a sampled classification task to reduce the classification error on a randomly sampled unseen classification task with a bilevel optimization scheme. This explicit learning aimed toward generalization allows the model to obtain a metric that separates well between unseen classes. We validate our L2G framework on mini-ImageNet and tiered-ImageNet datasets with two base meta-learning few-shot classification models, Prototypical Networks and Relation Networks. The results show that L2G significantly improves the performance of the two methods over episodic training. Further visualization shows that L2G obtains a metric space that clusters and separates unseen classes well.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/21/2018

A Meta-Learning Approach for Custom Model Training

Transfer-learning and meta-learning are two effective methods to apply k...
research
04/29/2020

Learning to Learn to Disambiguate: Meta-Learning for Few-Shot Word Sense Disambiguation

Deep learning methods typically rely on large amounts of annotated data ...
research
09/13/2019

Meta-Learning for Few-Shot Time Series Classification

Deep neural networks (DNNs) have achieved state-of-the-art results on ti...
research
12/19/2018

Modular meta-learning in abstract graph networks for combinatorial generalization

Modular meta-learning is a new framework that generalizes to unseen data...
research
09/29/2020

MetaMix: Improved Meta-Learning with Interpolation-based Consistency Regularization

Model-Agnostic Meta-Learning (MAML) and its variants are popular few-sho...
research
08/26/2021

Few-shot Visual Relationship Co-localization

In this paper, given a small bag of images, each containing a common but...
research
02/26/2022

Semantic Supervision: Enabling Generalization over Output Spaces

In this paper, we propose Semantic Supervision (SemSup) - a unified para...

Please sign up or login with your details

Forgot password? Click here to reset