Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning

12/05/2021
by   Zixuan Ke, et al.
20

Continual learning (CL) learns a sequence of tasks incrementally with the goal of achieving two main objectives: overcoming catastrophic forgetting (CF) and encouraging knowledge transfer (KT) across tasks. However, most existing techniques focus only on overcoming CF and have no mechanism to encourage KT, and thus do not do well in KT. Although several papers have tried to deal with both CF and KT, our experiments show that they suffer from serious CF when the tasks do not have much shared knowledge. Another observation is that most current CL methods do not use pre-trained models, but it has been shown that such models can significantly improve the end task performance. For example, in natural language processing, fine-tuning a BERT-like pre-trained language model is one of the most effective approaches. However, for CL, this approach suffers from serious CF. An interesting question is how to make the best use of pre-trained models for CL. This paper proposes a novel model called CTR to solve these problems. Our experimental results demonstrate the effectiveness of CTR

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2022

Continual Pre-Training Mitigates Forgetting in Language and Vision

Pre-trained models are nowadays a fundamental component of machine learn...
research
04/08/2020

CALM: Continuous Adaptive Learning for Language Modeling

Training large language representation models has become a standard in t...
research
02/05/2020

K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters

We study the problem of injecting knowledge into large pre-trained model...
research
05/19/2022

EXPANSE: A Deep Continual / Progressive Learning System for Deep Transfer Learning

Deep transfer learning techniques try to tackle the limitations of deep ...
research
02/22/2023

Preventing Catastrophic Forgetting in Continual Learning of New Natural Language Tasks

Multi-Task Learning (MTL) is widely-accepted in Natural Language Process...
research
06/22/2023

Class-Incremental Learning based on Label Generation

Despite the great success of pre-trained language models, it is still a ...
research
03/09/2022

Memory Efficient Continual Learning for Neural Text Classification

Learning text classifiers based on pre-trained language models has becom...

Please sign up or login with your details

Forgot password? Click here to reset