Continual Few-Shot Learning Using HyperTransformers

01/11/2023
by   Max Vladymyrov, et al.
0

We focus on the problem of learning without forgetting from multiple tasks arriving sequentially, where each task is defined using a few-shot episode of novel or already seen classes. We approach this problem using the recently published HyperTransformer (HT), a Transformer-based hypernetwork that generates specialized task-specific CNN weights directly from the support set. In order to learn from a continual sequence of tasks, we propose to recursively re-use the generated weights as input to the HT for the next task. This way, the generated CNN weights themselves act as a representation of previously learned tasks, and the HT is trained to update these weights so that the new task can be learned without forgetting past tasks. This approach is different from most continual learning algorithms that typically rely on using replay buffers, weight regularization or task-dependent architectural changes. We demonstrate that our proposed Continual HyperTransformer method equipped with a prototypical loss is capable of learning and retaining knowledge about past tasks for a variety of scenarios, including learning from mini-batches, and task-incremental and class-incremental learning scenarios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/26/2021

Dynamic VAEs with Generative Replay for Continual Zero-shot Learning

Continual zero-shot learning(CZSL) is a new domain to classify objects s...
research
04/19/2021

Few-shot Continual Learning: a Brain-inspired Approach

It is an important yet challenging setting to continually learn new task...
research
06/22/2018

Continuous Learning in Single-Incremental-Task Scenarios

It was recently shown that architectural, regularization and rehearsal s...
research
10/13/2021

Subspace Regularizers for Few-Shot Class Incremental Learning

Few-shot class incremental learning – the problem of updating a trained ...
research
11/21/2020

Learn to Bind and Grow Neural Structures

Task-incremental learning involves the challenging problem of learning n...
research
06/21/2023

TADIL: Task-Agnostic Domain-Incremental Learning through Task-ID Inference using Transformer Nearest-Centroid Embeddings

Machine Learning (ML) models struggle with data that changes over time o...
research
11/22/2022

Gated Class-Attention with Cascaded Feature Drift Compensation for Exemplar-free Continual Learning of Vision Transformers

In this paper we propose a new method for exemplar-free class incrementa...

Please sign up or login with your details

Forgot password? Click here to reset