A Unifying Bayesian View of Continual Learning

02/18/2019
by   Sebastian Farquhar, et al.
10

Some machine learning applications require continual learning - where data comes in a sequence of datasets, each is used for training and then permanently discarded. From a Bayesian perspective, continual learning seems straightforward: Given the model posterior one would simply use this as the prior for the next task. However, exact posterior evaluation is intractable with many models, especially with Bayesian neural networks (BNNs). Instead, posterior approximations are often sought. Unfortunately, when posterior approximations are used, prior-focused approaches do not succeed in evaluations designed to capture properties of realistic continual learning use cases. As an alternative to prior-focused methods, we introduce a new approximate Bayesian derivation of the continual learning loss. Our loss does not rely on the posterior from earlier tasks, and instead adapts the model itself by changing the likelihood term. We call these approaches likelihood-focused. We then combine prior- and likelihood-focused methods into one objective, tying the two views together under a single unifying framework of approximate Bayesian continual learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2018

Towards Robust Evaluations of Continual Learning

Continual learning experiments used in current deep learning papers do n...
research
11/11/2022

Understanding Approximation for Bayesian Inference in Neural Networks

Bayesian inference has theoretical attractions as a principled framework...
research
10/06/2020

Sequential Changepoint Detection in Neural Networks with Checkpoints

We introduce a framework for online changepoint detection and simultaneo...
research
03/01/2021

Posterior Meta-Replay for Continual Learning

Continual Learning (CL) algorithms have recently received a lot of atten...
research
07/11/2022

Repairing Neural Networks by Leaving the Right Past Behind

Prediction failures of machine learning models often arise from deficien...
research
06/10/2020

Continual Learning for Affective Computing

Real-world application require affect perception models to be sensitive ...
research
05/30/2023

Class Conditional Gaussians for Continual Learning

Dealing with representation shift is one of the main problems in online ...

Please sign up or login with your details

Forgot password? Click here to reset