Encoders and Ensembles for Task-Free Continual Learning

05/27/2021
by   Murray Shanahan, et al.
3

We present an architecture that is effective for continual learning in an especially demanding setting, where task boundaries do not exist or are unknown. Our architecture comprises an encoder, pre-trained on a separate dataset, and an ensemble of simple one-layer classifiers. Two main innovations are required to make this combination work. First, the provision of suitably generic pre-trained encoders has been made possible thanks to recent progress in self-supervised training methods. Second, pairing each classifier in the ensemble with a key, where the key-space is identical to the latent space of the encoder, allows them to be used collectively, yet selectively, via k-nearest neighbour lookup. We show that models trained with the encoders-and-ensembles architecture are state-of-the-art for the task-free setting on standard image classification continual learning benchmarks, and improve on prior state-of-the-art by a large margin in the most challenging cases. We also show that the architecture learns well in a fully incremental setting, where one class is learned at a time, and we demonstrate its effectiveness in this setting with up to 100 classes. Finally, we show that the architecture works in a task-free continual learning context where the data distribution changes gradually, and existing approaches requiring knowledge of task boundaries cannot be applied.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/24/2020

Energy-Based Models for Continual Learning

We motivate Energy-Based Models (EBMs) as a promising model class for co...
research
03/09/2023

SLCA: Slow Learner with Classifier Alignment for Continual Learning on a Pre-trained Model

The goal of continual learning is to improve the performance of recognit...
research
01/03/2020

A Neural Dirichlet Process Mixture Model for Task-Free Continual Learning

Despite the growing interest in continual learning, most of its contempo...
research
11/27/2022

Neural Architecture for Online Ensemble Continual Learning

Continual learning with an increasing number of classes is a challenging...
research
07/09/2023

Class-Incremental Mixture of Gaussians for Deep Continual Learning

Continual learning models for stationary data focus on learning and reta...
research
04/30/2022

Foundational Models for Continual Learning: An Empirical Study of Latent Replay

Rapid development of large-scale pre-training has resulted in foundation...
research
01/09/2023

CaSpeR: Latent Spectral Regularization for Continual Learning

While biological intelligence grows organically as new knowledge is gath...

Please sign up or login with your details

Forgot password? Click here to reset