Overcoming the Stability Gap in Continual Learning

06/02/2023
by   Md Yousuf Harun, et al.
0

In many real-world applications, deep neural networks are retrained from scratch as a dataset grows in size. Given the computational expense for retraining networks, it has been argued that continual learning could make updating networks more efficient. An obstacle to achieving this goal is the stability gap, which refers to an observation that when updating on new data, performance on previously learned data degrades before recovering. Addressing this problem would enable continual learning to learn new data with fewer network updates, resulting in increased computational efficiency. We study how to mitigate the stability gap in rehearsal (or experience replay), a widely employed continual learning method. We test a variety of hypotheses to understand why the stability gap occurs. This leads us to discover a method that vastly reduces this gap. In experiments on a large-scale incremental class learning setting, we are able to significantly reduce the number of network updates to recover performance. Our work has the potential to advance the state-of-the-art in continual learning for real-world applications along with reducing the carbon footprint required to maintain updated neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/19/2023

SIESTA: Efficient Online Continual Learning with Sleep

In supervised continual learning, a deep neural network (DNN) is updated...
research
07/11/2020

Batch-level Experience Replay with Review for Continual Learning

Continual learning is a branch of deep learning that seeks to strike a b...
research
02/14/2022

Design of Explainability Module with Experts in the Loop for Visualization and Dynamic Adjustment of Continual Learning

Continual learning can enable neural networks to evolve by learning new ...
research
05/19/2023

Conditional Online Learning for Keyword Spotting

Modern approaches for keyword spotting rely on training deep neural netw...
research
06/10/2020

Dataset Condensation with Gradient Matching

Efficient training of deep neural networks is an increasingly important ...
research
05/28/2018

Parallel Weight Consolidation: A Brain Segmentation Case Study

Collecting the large datasets needed to train deep neural networks can b...
research
03/10/2023

Lifelong Machine Learning Potentials

Machine learning potentials (MLPs) trained on accurate quantum chemical ...

Please sign up or login with your details

Forgot password? Click here to reset