Cold Start Streaming Learning for Deep Networks

11/09/2022
by   Cameron R. Wolfe, et al.
0

The ability to dynamically adapt neural networks to newly-available data without performance deterioration would revolutionize deep learning applications. Streaming learning (i.e., learning from one data example at a time) has the potential to enable such real-time adaptation, but current approaches i) freeze a majority of network parameters during streaming and ii) are dependent upon offline, base initialization procedures over large subsets of data, which damages performance and limits applicability. To mitigate these shortcomings, we propose Cold Start Streaming Learning (CSSL), a simple, end-to-end approach for streaming learning with deep networks that uses a combination of replay and data augmentation to avoid catastrophic forgetting. Because CSSL updates all model parameters during streaming, the algorithm is capable of beginning streaming from a random initialization, making base initialization optional. Going further, the algorithm's simplicity allows theoretical convergence guarantees to be derived using analysis of the Neural Tangent Random Feature (NTRF). In experiments, we find that CSSL outperforms existing baselines for streaming learning in experiments on CIFAR100, ImageNet, and Core50 datasets. Additionally, we propose a novel multi-task streaming learning setting and show that CSSL performs favorably in this domain. Put simply, CSSL performs well and demonstrates that the complicated, multi-step training pipelines adopted by most streaming methodologies can be replaced with a simple, end-to-end learning approach without sacrificing performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/04/2019

Lifelong Machine Learning with Deep Streaming Linear Discriminant Analysis

When a robot acquires new information, ideally it would immediately be c...
research
09/15/2023

Continual Learning with Deep Streaming Regularized Discriminant Analysis

Continual learning is increasingly sought after in real world machine le...
research
09/15/2021

UniST: Unified End-to-end Model for Streaming and Non-streaming Speech Translation

This paper presents a unified end-to-end frame-work for both streaming a...
research
12/21/2022

Incremental Learning for Neural Radiance Field with Uncertainty-Filtered Knowledge Distillation

Recent neural radiance field (NeRF) representation has achieved great su...
research
10/20/2021

Class Incremental Online Streaming Learning

A wide variety of methods have been developed to enable lifelong learnin...
research
06/17/2016

DecomposeMe: Simplifying ConvNets for End-to-End Learning

Deep learning and convolutional neural networks (ConvNets) have been suc...
research
06/24/2019

Streaming Adaptation of Deep Forecasting Models using Adaptive Recurrent Units

We present ARU, an Adaptive Recurrent Unit for streaming adaptation of d...

Please sign up or login with your details

Forgot password? Click here to reset