Memory Efficient Meta-Learning with Large Images

07/02/2021
by   John Bronskill, et al.
10

Meta learning approaches to few-shot classification are computationally efficient at test time requiring just a few optimization steps or single forward pass to learn a new task, but they remain highly memory-intensive to train. This limitation arises because a task's entire support set, which can contain up to 1000 images, must be processed before an optimization step can be taken. Harnessing the performance gains offered by large images thus requires either parallelizing the meta-learner across multiple GPUs, which may not be available, or trade-offs between task and image size when memory constraints apply. We improve on both options by proposing LITE, a general and memory efficient episodic training scheme that enables meta-training on large tasks composed of large images on a single GPU. We achieve this by observing that the gradients for a task can be decomposed into a sum of gradients over the task's training images. This enables us to perform a forward pass on a task's entire training set but realize significant memory savings by back-propagating only a random subset of these images which we show is an unbiased approximation of the full gradient. We use LITE to train meta-learners and demonstrate new state-of-the-art accuracy on the real-world ORBIT benchmark and 3 of the 4 parts of the challenging VTAB+MD benchmark relative to leading meta-learners. LITE also enables meta-learners to be competitive with transfer learning approaches but at a fraction of the test-time computational cost, thus serving as a counterpoint to the recent narrative that transfer learning is all you need for few-shot classification.

READ FULL TEXT
research
04/06/2021

Comparing Transfer and Meta Learning Approaches on a Unified Few-Shot Classification Benchmark

Meta and transfer learning are two successful families of approaches to ...
research
02/11/2020

Incremental Learning for Metric-Based Meta-Learners

Majority of the modern meta-learning methods for few-shot classification...
research
04/11/2019

MxML: Mixture of Meta-Learners for Few-Shot Classification

A meta-model is trained on a distribution of similar tasks such that it ...
research
07/15/2021

A Channel Coding Benchmark for Meta-Learning

Meta-learning provides a popular and effective family of methods for dat...
research
06/20/2022

Contextual Squeeze-and-Excitation for Efficient Few-Shot Image Classification

Recent years have seen a growth in user-centric applications that requir...
research
06/08/2020

Multi-step Estimation for Gradient-based Meta-learning

Gradient-based meta-learning approaches have been successful in few-shot...
research
10/21/2021

On Hard Episodes in Meta-Learning

Existing meta-learners primarily focus on improving the average task acc...

Please sign up or login with your details

Forgot password? Click here to reset