Meta-Regularization by Enforcing Mutual-Exclusiveness

01/24/2021
by   Edwin Pan, et al.
0

Meta-learning models have two objectives. First, they need to be able to make predictions over a range of task distributions while utilizing only a small amount of training data. Second, they also need to adapt to new novel unseen tasks at meta-test time again by using only a small amount of training data from that task. It is the second objective where meta-learning models fail for non-mutually exclusive tasks due to task overfitting. Given that guaranteeing mutually exclusive tasks is often difficult, there is a significant need for regularization methods that can help reduce the impact of task-memorization in meta-learning. For example, in the case of N-way, K-shot classification problems, tasks becomes non-mutually exclusive when the labels associated with each task is fixed. Under this design, the model will simply memorize the class labels of all the training tasks, and thus will fail to recognize a new task (class) at meta-test time. A direct observable consequence of this memorization is that the meta-learning model simply ignores the task-specific training data in favor of directly classifying based on the test-data input. In our work, we propose a regularization technique for meta-learning models that gives the model designer more control over the information flow during meta-training. Our method consists of a regularization function that is constructed by maximizing the distance between task-summary statistics, in the case of black-box models and task specific network parameters in the case of optimization based models during meta-training. Our proposed regularization function shows an accuracy boost of ∼ 36% on the Omniglot dataset for 5-way, 1-shot classification using black-box method and for 20-way, 1-shot classification problem using optimization-based method.

READ FULL TEXT

page 4

page 8

research
12/09/2019

Meta-Learning without Memorization

The ability to learn new concepts with small amounts of data is a critic...
research
10/02/2021

An Optimization-Based Meta-Learning Model for MRI Reconstruction with Diverse Dataset

Purpose: This work aims at developing a generalizable MRI reconstruction...
research
02/22/2021

Unsupervised Meta Learning for One Shot Title Compression in Voice Commerce

Product title compression for voice and mobile commerce is a well studie...
research
05/28/2019

Discrete Infomax Codes for Meta-Learning

Learning compact discrete representations of data is itself a key task i...
research
03/06/2020

Meta-SVDD: Probabilistic Meta-Learning for One-Class Classification in Cancer Histology Images

To train a robust deep learning model, one usually needs a balanced set ...
research
03/29/2022

Zero-shot meta-learning for small-scale data from human subjects

While developments in machine learning led to impressive performance gai...
research
03/15/2021

How to distribute data across tasks for meta-learning?

Meta-learning models transfer the knowledge acquired from previous tasks...

Please sign up or login with your details

Forgot password? Click here to reset