TaskNorm: Rethinking Batch Normalization for Meta-Learning

03/06/2020
by   John Bronskill, et al.
12

Modern meta-learning approaches for image classification rely on increasingly deep networks to achieve state-of-the-art performance, making batch normalization an essential component of meta-learning pipelines. However, the hierarchical nature of the meta-learning setting presents several challenges that can render conventional batch normalization ineffective, giving rise to the need to rethink normalization in this setting. We evaluate a range of approaches to batch normalization for meta-learning scenarios, and develop a novel approach that we call TaskNorm. Experiments on fourteen datasets demonstrate that the choice of batch normalization has a dramatic effect on both classification accuracy and training time for both gradient based and gradient-free meta-learning approaches. Importantly, TaskNorm is found to consistently improve performance. Finally, we provide a set of best practices for normalization that will allow fair comparison of meta-learning algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/17/2020

Unraveling Meta-Learning: Understanding Feature Representations for Few-Shot Tasks

Meta-learning algorithms produce feature extractors which achieve state-...
research
02/27/2020

Is the Meta-Learning Idea Able to Improve the Generalization of Deep Neural Networks on the Standard Supervised Learning?

Substantial efforts have been made on improving the generalization abili...
research
10/08/2018

Meta-Learning: A Survey

Meta-learning, or learning to learn, is the science of systematically ob...
research
08/05/2023

Meta-learning in healthcare: A survey

As a subset of machine learning, meta-learning, or learning to learn, ai...
research
08/02/2022

Stochastic Deep Networks with Linear Competing Units for Model-Agnostic Meta-Learning

This work addresses meta-learning (ML) by considering deep networks with...
research
04/02/2022

AutoProtoNet: Interpretability for Prototypical Networks

In meta-learning approaches, it is difficult for a practitioner to make ...
research
04/30/2021

Faster Meta Update Strategy for Noise-Robust Deep Learning

It has been shown that deep neural networks are prone to overfitting on ...

Please sign up or login with your details

Forgot password? Click here to reset