Task Agnostic Representation Consolidation: a Self-supervised based Continual Learning Approach

07/13/2022
by   Prashant Bhat, et al.
28

Continual learning (CL) over non-stationary data streams remains one of the long-standing challenges in deep neural networks (DNNs) as they are prone to catastrophic forgetting. CL models can benefit from self-supervised pre-training as it enables learning more generalizable task-agnostic features. However, the effect of self-supervised pre-training diminishes as the length of task sequences increases. Furthermore, the domain shift between pre-training data distribution and the task distribution reduces the generalizability of the learned representations. To address these limitations, we propose Task Agnostic Representation Consolidation (TARC), a two-stage training paradigm for CL that intertwines task-agnostic and task-specific learning whereby self-supervised training is followed by supervised learning for each task. To further restrict the deviation from the learned representations in the self-supervised stage, we employ a task-agnostic auxiliary loss during the supervised stage. We show that our training paradigm can be easily added to memory- or regularization-based approaches and provides consistent performance gain across more challenging CL settings. We further show that it leads to more robust and well-calibrated models.

READ FULL TEXT

page 7

page 8

page 14

research
03/25/2021

Self-Supervised Training Enhances Online Continual Learning

In continual learning, a system must incrementally learn from a non-stat...
research
05/19/2022

Continual Pre-Training Mitigates Forgetting in Language and Vision

Pre-trained models are nowadays a fundamental component of machine learn...
research
08/25/2021

Learning From Long-Tailed Data With Noisy Labels

Class imbalance and noisy labels are the norm rather than the exception ...
research
03/31/2020

Towards Lifelong Self-Supervision For Unpaired Image-to-Image Translation

Unpaired Image-to-Image Translation (I2IT) tasks often suffer from lack ...
research
07/22/2022

Discrete Key-Value Bottleneck

Deep neural networks perform well on prediction and classification tasks...
research
03/29/2022

Learning neural audio features without supervision

Deep audio classification, traditionally cast as training a deep neural ...
research
01/07/2021

Contextual Classification Using Self-Supervised Auxiliary Models for Deep Neural Networks

Classification problems solved with deep neural networks (DNNs) typicall...

Please sign up or login with your details

Forgot password? Click here to reset