Lifelong Generative Modelling Using Dynamic Expansion Graph Model

12/15/2021
by   Fei Ye, et al.
0

Variational Autoencoders (VAEs) suffer from degenerated performance, when learning several successive tasks. This is caused by catastrophic forgetting. In order to address the knowledge loss, VAEs are using either Generative Replay (GR) mechanisms or Expanding Network Architectures (ENA). In this paper we study the forgetting behaviour of VAEs using a joint GR and ENA methodology, by deriving an upper bound on the negative marginal log-likelihood. This theoretical analysis provides new insights into how VAEs forget the previously learnt knowledge during lifelong learning. The analysis indicates the best performance achieved when considering model mixtures, under the ENA framework, where there are no restrictions on the number of components. However, an ENA-based approach may require an excessive number of parameters. This motivates us to propose a novel Dynamic Expansion Graph Model (DEGM). DEGM expands its architecture, according to the novelty associated with each new databases, when compared to the information already learnt by the network from previous tasks. DEGM training optimizes knowledge structuring, characterizing the joint probabilistic representations corresponding to the past and more recently learned tasks. We demonstrate that DEGM guarantees optimal performance for each task while also minimizing the required number of parameters. Supplementary materials (SM) and source code are available in https://github.com/dtuzi123/Expansion-Graph-Model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/25/2022

Supplemental Material: Lifelong Generative Modelling Using Dynamic Expansion Graph Model

In this article, we provide the appendix for Lifelong Generative Modelli...
research
08/25/2021

Lifelong Infinite Mixture Model Based on Knowledge-Driven Dirichlet Process

Recent research efforts in lifelong learning propose to grow a mixture o...
research
11/22/2021

DyTox: Transformers for Continual Learning with DYnamic TOken eXpansion

Deep network architectures struggle to continually learn new tasks witho...
research
07/09/2021

Lifelong Teacher-Student Network Learning

A unique cognitive capability of humans consists in their ability to acq...
research
10/31/2022

Reduce Catastrophic Forgetting of Dense Retrieval Training with Teleportation Negatives

In this paper, we investigate the instability in the standard dense retr...
research
10/12/2022

Task-Free Continual Learning via Online Discrepancy Distance Learning

Learning from non-stationary data streams, also called Task-Free Continu...
research
10/26/2022

Characterizing Datapoints via Second-Split Forgetting

Researchers investigating example hardness have increasingly focused on ...

Please sign up or login with your details

Forgot password? Click here to reset