Enabling Continual Learning with Differentiable Hebbian Plasticity

06/30/2020
by   Vithursan Thangarasa, et al.
0

Continual learning is the problem of sequentially learning new tasks or knowledge while protecting previously acquired knowledge. However, catastrophic forgetting poses a grand challenge for neural networks performing such learning process. Thus, neural networks that are deployed in the real world often struggle in scenarios where the data distribution is non-stationary (concept drift), imbalanced, or not always fully available, i.e., rare edge cases. We propose a Differentiable Hebbian Consolidation model which is composed of a Differentiable Hebbian Plasticity (DHP) Softmax layer that adds a rapid learning plastic component (compressed episodic memory) to the fixed (slow changing) parameters of the softmax output layer; enabling learned representations to be retained for a longer timescale. We demonstrate the flexibility of our method by integrating well-known task-specific synaptic consolidation methods to penalize changes in the slow weights that are important for each target task. We evaluate our approach on the Permuted MNIST, Split MNIST and Vision Datasets Mixture benchmarks, and introduce an imbalanced variant of Permuted MNIST – a dataset that combines the challenges of class imbalance and concept drift. Our proposed model requires no additional hyperparameters and outperforms comparable baselines by reducing forgetting.

READ FULL TEXT

page 1

page 4

page 5

research
06/24/2021

Task-agnostic Continual Learning with Hybrid Probabilistic Models

Learning new tasks continuously without forgetting on a constantly chang...
research
05/17/2021

Shared and Private VAEs with Generative Replay for Continual Learning

Continual learning tries to learn new tasks without forgetting previousl...
research
10/25/2021

Mixture-of-Variational-Experts for Continual Learning

One significant shortcoming of machine learning is the poor ability of m...
research
06/03/2021

Continual Learning in Deep Networks: an Analysis of the Last Layer

We study how different output layer types of a deep neural network learn...
research
12/12/2022

Learning on non-stationary data with re-weighting

Many real-world learning scenarios face the challenge of slow concept dr...
research
08/26/2021

Continual learning under domain transfer with sparse synaptic bursting

Existing machines are functionally specific tools that were made for eas...
research
05/18/2021

DRILL: Dynamic Representations for Imbalanced Lifelong Learning

Continual or lifelong learning has been a long-standing challenge in mac...

Please sign up or login with your details

Forgot password? Click here to reset