Overcoming Catastrophic Forgetting by Generative Regularization

12/03/2019
by   Patrick H. Chen, et al.
0

In this paper, we propose a new method to overcome catastrophic forgetting by adding generative regularization to Bayesian inference framework. We could construct generative regularization term for all given models by leveraging Energy-based models and Langevin-Dynamic sampling. By combining discriminative and generative loss together, we show that this intuitively provides a better posterior formulation in Bayesian inference. Experimental results show that the proposed method outperforms state of-the-art methods on a variety of tasks, avoiding catastrophic forgetting in continual learning. In particular, the proposed method outperforms previous methos over 10% in Fashion-MNIST dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/04/2023

On Sequential Bayesian Inference for Continual Learning

Sequential Bayesian inference can be used for continual learning to prev...
research
01/15/2021

Learning Invariant Representation for Continual Learning

Continual learning aims to provide intelligent agents that are capable o...
research
03/31/2019

Learn to Grow: A Continual Structure Learning Framework for Overcoming Catastrophic Forgetting

Addressing catastrophic forgetting is one of the key challenges in conti...
research
04/11/2023

Task Difficulty Aware Parameter Allocation Regularization for Lifelong Learning

Parameter regularization or allocation methods are effective in overcomi...
research
03/19/2020

Lifelong Learning with Searchable Extension Units

Lifelong learning remains an open problem. One of its main difficulties ...
research
04/17/2021

Lifelong Learning with Sketched Structural Regularization

Preventing catastrophic forgetting while continually learning new tasks ...
research
02/11/2022

Continual Learning with Invertible Generative Models

Catastrophic forgetting (CF) happens whenever a neural network overwrite...

Please sign up or login with your details

Forgot password? Click here to reset