Catastrophic Forgetting in the Context of Model Updates

06/16/2023
by   Rich Harang, et al.
0

A large obstacle to deploying deep learning models in practice is the process of updating models post-deployment (ideally, frequently). Deep neural networks can cost many thousands of dollars to train. When new data comes in the pipeline, you can train a new model from scratch (randomly initialized weights) on all existing data. Instead, you can take an existing model and fine-tune (continue to train) it on new data. The former is costly and slow. The latter is cheap and fast, but catastrophic forgetting generally causes the new model to 'forget' how to classify older data well. There are a plethora of complicated techniques to keep models from forgetting their past learnings. Arguably the most basic is to mix in a small amount of past data into the new data during fine-tuning: also known as 'data rehearsal'. In this paper, we compare various methods of limiting catastrophic forgetting and conclude that if you can maintain access to a portion of your past data (or tasks), data rehearsal is ideal in terms of overall accuracy across all time periods, and performs even better when combined with methods like Elastic Weight Consolidation (EWC). Especially when the amount of past data (past 'tasks') is large compared to new data, the cost of updating an existing model is far cheaper and faster than training a new model from scratch.

READ FULL TEXT
research
05/04/2022

Explain to Not Forget: Defending Against Catastrophic Forgetting with XAI

The ability to continuously process and retain new information like we d...
research
05/16/2023

Towards Lifelong Learning for Software Analytics Models: Empirical Study on Brown Build and Risk Prediction

Nowadays, software analytics tools using machine learning (ML) models to...
research
02/21/2022

BERT WEAVER: Using WEight AVERaging to Enable Lifelong Learning for Transformer-based Models

Recent developments in transfer learning have boosted the advancements i...
research
01/05/2021

One vs Previous and Similar Classes Learning – A Comparative Study

When dealing with multi-class classification problems, it is common prac...
research
04/29/2020

Neural Network Retraining for Model Serving

We propose incremental (re)training of a neural network model to cope wi...
research
03/27/2021

Addressing catastrophic forgetting for medical domain expansion

Model brittleness is a key concern when deploying deep learning models i...
research
06/29/2016

Learning without Forgetting

When building a unified vision system or gradually adding new capabiliti...

Please sign up or login with your details

Forgot password? Click here to reset