Learning to Remember from a Multi-Task Teacher

10/10/2019
by   Yuwen Xiong, et al.
0

Recent studies on catastrophic forgetting during sequential learning typically focus on fixing the accuracy of the predictions for a previously learned task. In this paper we argue that the outputs of neural networks are subject to rapid changes when learning a new data distribution, and networks that appear to "forget" everything still contain useful representation towards previous tasks. Instead of enforcing the output accuracy to stay the same, we propose to reduce the effect of catastrophic forgetting on the representation level, as the output layer can be quickly recovered later with a small number of examples. Towards this goal, we propose an experimental setup that measures the amount of representational forgetting, and develop a novel meta-learning algorithm to overcome this issue. The proposed meta-learner produces weight updates of a sequential learning network, mimicking a multi-task teacher network's representation. We show that our meta-learner can improve its learned representations on new tasks, while maintaining a good representation for old tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/09/2019

Meta-learnt priors slow down catastrophic forgetting in neural networks

Current training regimes for deep learning usually involve exposure to a...
research
04/30/2020

Bayesian Online Meta-Learning with Laplace Approximation

Neural networks are known to suffer from catastrophic forgetting when tr...
research
03/20/2019

Regularize, Expand and Compress: Multi-task based Lifelong Learning via NonExpansive AutoML

Lifelong learning, the problem of continual learning where tasks arrive ...
research
05/04/2022

Explain to Not Forget: Defending Against Catastrophic Forgetting with XAI

The ability to continuously process and retain new information like we d...
research
11/15/2019

Forgetting to learn logic programs

Most program induction approaches require predefined, often hand-enginee...
research
11/11/2021

Kronecker Factorization for Preventing Catastrophic Forgetting in Large-scale Medical Entity Linking

Multi-task learning is useful in NLP because it is often practically des...
research
09/23/2021

The Role of Bio-Inspired Modularity in General Learning

One goal of general intelligence is to learn novel information without o...

Please sign up or login with your details

Forgot password? Click here to reset