On Quadratic Penalties in Elastic Weight Consolidation

12/11/2017
by   Ferenc Huszár, et al.
0

Elastic weight consolidation (EWC, Kirkpatrick et al, 2017) is a novel algorithm designed to safeguard against catastrophic forgetting in neural networks. EWC can be seen as an approximation to Laplace propagation (Eskin et al, 2004), and this view is consistent with the motivation given by Kirkpatrick et al (2017). In this note, I present an extended derivation that covers the case when there are more than two tasks. I show that the quadratic penalties in EWC are inconsistent with this derivation and might lead to double-counting data from earlier tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/09/2017

Improved Space-efficient Linear Time Algorithms for Some Classical Graph Problems

This short note provides space-efficient linear time algorithms for comp...
research
09/25/2019

Towards continuous learning for glioma segmentation with elastic weight consolidation

When finetuning a convolutional neural network (CNN) on data from a new ...
research
01/31/2023

The weight spectrum of several families of Reed-Muller codes

We determine the weight spectrum of RM(m-3,m) for m≥ 6, of RM(m-4,m) for...
research
10/23/2019

Double-Counting Problem of the Bonus-Malus System

The bonus-malus system (BMS) is a widely used premium adjustment mechani...
research
06/07/2022

The b-symbol weight distributions of all semiprimitive irreducible cyclic codes

Up to a new invariant μ(b), the complete b-symbol weight distribution of...
research
05/25/2019

Cold Case: The Lost MNIST Digits

Although the popular MNIST dataset [LeCun et al., 1994] is derived from ...
research
01/28/2022

Discriminating Defense Against DDoS Attacks; a Novel Approach

A recent paper (circa 2020) by Osterwile et al., entitled "21 Years of D...

Please sign up or login with your details

Forgot password? Click here to reset