DLCFT: Deep Linear Continual Fine-Tuning for General Incremental Learning

08/17/2022
by   Hyounguk Shon, et al.
0

Pre-trained representation is one of the key elements in the success of modern deep learning. However, existing works on continual learning methods have mostly focused on learning models incrementally from scratch. In this paper, we explore an alternative framework to incremental learning where we continually fine-tune the model from a pre-trained representation. Our method takes advantage of linearization technique of a pre-trained neural network for simple and effective continual learning. We show that this allows us to design a linear model where quadratic parameter regularization method is placed as the optimal continual learning policy, and at the same time enjoying the high performance of neural networks. We also show that the proposed algorithm enables parameter regularization methods to be applied to class-incremental problems. Additionally, we provide a theoretical reason why the existing parameter-space regularization algorithms such as EWC underperform on neural networks trained with cross-entropy loss. We show that the proposed method can prevent forgetting while achieving high continual fine-tuning performance on image classification tasks. To show that our method can be applied to general continual learning settings, we evaluate our method in data-incremental, task-incremental, and class-incremental learning problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2023

Continual Learning with Pretrained Backbones by Tuning in the Input Space

The intrinsic difficulty in adapting deep learning models to non-station...
research
01/23/2022

vCLIMB: A Novel Video Class Incremental Learning Benchmark

Continual learning (CL) is under-explored in the video domain. The few e...
research
02/28/2023

Adapter Incremental Continual Learning of Efficient Audio Spectrogram Transformers

Continual learning involves training neural networks incrementally for n...
research
07/16/2023

Tangent Model Composition for Ensembling and Continual Fine-tuning

Tangent Model Composition (TMC) is a method to combine component models ...
research
09/01/2022

Incremental Online Learning Algorithms Comparison for Gesture and Visual Smart Sensors

Tiny machine learning (TinyML) in IoT systems exploits MCUs as edge devi...
research
03/10/2023

Lifelong Machine Learning Potentials

Machine learning potentials (MLPs) trained on accurate quantum chemical ...
research
11/22/2021

FFNB: Forgetting-Free Neural Blocks for Deep Continual Visual Learning

Deep neural networks (DNNs) have recently achieved a great success in co...

Please sign up or login with your details

Forgot password? Click here to reset