Adapt Your Teacher: Improving Knowledge Distillation for Exemplar-free Continual Learning

08/18/2023
by   Filip Szatkowski, et al.
0

In this work, we investigate exemplar-free class incremental learning (CIL) with knowledge distillation (KD) as a regularization strategy, aiming to prevent forgetting. KD-based methods are successfully used in CIL, but they often struggle to regularize the model without access to exemplars of the training data from previous tasks. Our analysis reveals that this issue originates from substantial representation shifts in the teacher network when dealing with out-of-distribution data. This causes large errors in the KD loss component, leading to performance degradation in CIL. Inspired by recent test-time adaptation methods, we introduce Teacher Adaptation (TA), a method that concurrently updates the teacher and the main model during incremental training. Our method seamlessly integrates with KD-based CIL approaches and allows for consistent enhancement of their performance across multiple exemplar-free CIL benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/31/2022

A Closer Look at Rehearsal-Free Continual Learning

Continual learning describes a setting where machine learning models lea...
research
06/01/2023

Teacher Agent: A Non-Knowledge Distillation Method for Rehearsal-based Video Incremental Learning

With the rise in popularity of video-based social media, new categories ...
research
12/10/2022

LEAD: Liberal Feature-based Distillation for Dense Retrieval

Knowledge distillation is often used to transfer knowledge from a strong...
research
05/23/2022

Self-distilled Knowledge Delegator for Exemplar-free Class Incremental Learning

Exemplar-free incremental learning is extremely challenging due to inacc...
research
02/23/2022

Multi-Teacher Knowledge Distillation for Incremental Implicitly-Refined Classification

Incremental learning methods can learn new classes continually by distil...
research
04/04/2022

Re-examining Distillation For Continual Object Detection

Training models continually to detect and classify objects, from new cla...
research
03/10/2023

Dynamic Y-KD: A Hybrid Approach to Continual Instance Segmentation

Despite the success of deep learning methods on instance segmentation, t...

Please sign up or login with your details

Forgot password? Click here to reset