Bayesian Nonparametric Weight Factorization for Continual Learning

04/21/2020
by   Nikhil Mehta, et al.
0

Naively trained neural networks tend to experience catastrophic forgetting in sequential task settings, where data from previous tasks are unavailable. A number of methods, using various model expansion strategies, have been proposed recently as possible solutions. However, determining how much to expand the model is left to the practitioner, and typically a constant schedule is chosen for simplicity, regardless of how complex the incoming task is. Instead, we propose a principled Bayesian nonparametric approach based on the Indian Buffet Process (IBP) prior, letting the data determine how much to expand the model complexity. We pair this with a factorization of the neural network's weight matrices. Such an approach allows us to scale the number of factors of each weight matrix to the complexity of the task, while the IBP prior imposes weight factor sparsity and encourages factor reuse, promoting positive knowledge transfer between tasks. We demonstrate the effectiveness of our method on a number of continual learning benchmarks and analyze how weight factors are allocated and reused throughout the training.

READ FULL TEXT

page 2

page 10

research
01/03/2020

A Neural Dirichlet Process Mixture Model for Task-Free Continual Learning

Despite the growing interest in continual learning, most of its contempo...
research
01/04/2023

On Sequential Bayesian Inference for Continual Learning

Sequential Bayesian inference can be used for continual learning to prev...
research
10/15/2019

Compacting, Picking and Growing for Unforgetting Continual Learning

Continual lifelong learning is essential to many applications. In this p...
research
12/08/2019

Nonparametric Bayesian Structure Adaptation for Continual Learning

Continual Learning is a learning paradigm where machine learning models ...
research
03/11/2019

Continual Learning via Neural Pruning

We introduce Continual Learning via Neural Pruning (CLNP), a new method ...
research
12/06/2022

Statistical mechanics of continual learning: variational principle and mean-field potential

An obstacle to artificial general intelligence is set by the continual l...

Please sign up or login with your details

Forgot password? Click here to reset