Adaptively Integrated Knowledge Distillation and Prediction Uncertainty for Continual Learning

01/18/2023
by   Kanghao Chen, et al.
8

Current deep learning models often suffer from catastrophic forgetting of old knowledge when continually learning new knowledge. Existing strategies to alleviate this issue often fix the trade-off between keeping old knowledge (stability) and learning new knowledge (plasticity). However, the stability-plasticity trade-off during continual learning may need to be dynamically changed for better model performance. In this paper, we propose two novel ways to adaptively balance model stability and plasticity. The first one is to adaptively integrate multiple levels of old knowledge and transfer it to each block level in the new model. The second one uses prediction uncertainty of old knowledge to naturally tune the importance of learning new knowledge during model training. To our best knowledge, this is the first time to connect model prediction uncertainty and knowledge distillation for continual learning. In addition, this paper applies a modified CutMix particularly to augment the data for old knowledge, further alleviating the catastrophic forgetting issue. Extensive evaluations on the CIFAR100 and the ImageNet datasets confirmed the effectiveness of the proposed method for continual learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/26/2023

Preserving Linear Separability in Continual Learning by Backward Feature Projection

Catastrophic forgetting has been a major challenge in continual learning...
research
05/09/2023

SRIL: Selective Regularization for Class-Incremental Learning

Human intelligence gradually accepts new information and accumulates kno...
research
07/22/2023

Revisiting Distillation for Continual Learning on Visual Question Localized-Answering in Robotic Surgery

The visual-question localized-answering (VQLA) system can serve as a kno...
research
03/16/2023

Achieving a Better Stability-Plasticity Trade-off via Auxiliary Networks in Continual Learning

In contrast to the natural capabilities of humans to learn new tasks in ...
research
09/12/2022

Online Continual Learning via the Meta-learning Update with Multi-scale Knowledge Distillation and Data Augmentation

Continual learning aims to rapidly and continually learn the current tas...
research
03/10/2023

Dynamic Y-KD: A Hybrid Approach to Continual Instance Segmentation

Despite the success of deep learning methods on instance segmentation, t...
research
02/19/2021

Continual Learning for Blind Image Quality Assessment

The explosive growth of image data facilitates the fast development of i...

Please sign up or login with your details

Forgot password? Click here to reset