Anatomy of Catastrophic Forgetting: Hidden Representations and Task Semantics

07/14/2020
by   Vinay V. Ramasesh, et al.
0

A central challenge in developing versatile machine learning systems is catastrophic forgetting: a model trained on tasks in sequence will suffer significant performance drops on earlier tasks. Despite the ubiquity of catastrophic forgetting, there is limited understanding of the underlying process and its causes. In this paper, we address this important knowledge gap, investigating how forgetting affects representations in neural network models. Through representational analysis techniques, we find that deeper layers are disproportionately the source of forgetting. Supporting this, a study of methods to mitigate forgetting illustrates that they act to stabilize deeper layers. These insights enable the development of an analytic argument and empirical picture relating the degree of forgetting to representational similarity between tasks. Consistent with this picture, we observe maximal forgetting occurs for task sequences with intermediate similarity. We perform empirical studies on the standard split CIFAR-10 setup and also introduce a novel CIFAR-100 based task approximating realistic input distribution shift.

READ FULL TEXT

page 20

page 21

research
07/09/2021

Continual Learning in the Teacher-Student Setup: Impact of Task Similarity

Continual learning-the ability to learn many tasks in sequence-is critic...
research
09/16/2022

Causes of Catastrophic Forgetting in Class-Incremental Semantic Segmentation

Class-incremental learning for semantic segmentation (CiSS) is presently...
research
01/17/2022

Evaluating Inexact Unlearning Requires Revisiting Forgetting

Existing works in inexact machine unlearning focus on achieving indistin...
research
10/20/2021

Behavioral Experiments for Understanding Catastrophic Forgetting

In this paper we explore whether the fundamental tool of experimental ps...
research
11/11/2021

Lifelong Learning from Event-based Data

Lifelong learning is a long-standing aim for artificial agents that act ...
research
11/16/2018

On Training Recurrent Neural Networks for Lifelong Learning

Capacity saturation and catastrophic forgetting are the central challeng...
research
04/27/2020

A general approach to progressive learning

In biological learning, data is used to improve performance on the task ...

Please sign up or login with your details

Forgot password? Click here to reset