Provable and Efficient Continual Representation Learning

03/03/2022
by   Yingcong Li, et al.
0

In continual learning (CL), the goal is to design models that can learn a sequence of tasks without catastrophic forgetting. While there is a rich set of techniques for CL, relatively little understanding exists on how representations built by previous tasks benefit new tasks that are added to the network. To address this, we study the problem of continual representation learning (CRL) where we learn an evolving representation as new tasks arrive. Focusing on zero-forgetting methods where tasks are embedded in subnetworks (e.g., PackNet), we first provide experiments demonstrating CRL can significantly boost sample efficiency when learning new tasks. To explain this, we establish theoretical guarantees for CRL by providing sample complexity and generalization error bounds for new tasks by formalizing the statistical benefits of previously-learned representations. Our analysis and experiments also highlight the importance of the order in which we learn the tasks. Specifically, we show that CL benefits if the initial tasks have large sample size and high "representation diversity". Diversity ensures that adding new tasks incurs small representation mismatch and can be learned with few samples while training only few additional nonzero weights. Finally, we ask whether one can ensure each task subnetwork to be efficient during inference time while retaining the benefits of representation learning. To this end, we propose an inference-efficient variation of PackNet called Efficient Sparse PackNet (ESPN) which employs joint channel weight pruning. ESPN embeds tasks in channel-sparse subnets requiring up to 80 approximately retaining accuracy and is very competitive with a variety of baselines. In summary, this work takes a step towards data and compute-efficient CL with a representation learning perspective. GitHub page: https://github.com/ucr-optml/CtRL

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/06/2020

Disentangle-based Continual Graph Representation Learning

Graph embedding (GE) methods embed nodes (and/or edges) in graph into a ...
research
05/29/2019

Meta-Learning Representations for Continual Learning

A continual learning agent should be able to build on top of existing kn...
research
06/01/2023

Task Relation-aware Continual User Representation Learning

User modeling, which learns to represent users into a low-dimensional re...
research
02/07/2023

Utility-based Perturbed Gradient Descent: An Optimizer for Continual Learning

Modern representation learning methods may fail to adapt quickly under n...
research
10/04/2020

Remembering for the Right Reasons: Explanations Reduce Catastrophic Forgetting

The goal of continual learning (CL) is to learn a sequence of tasks with...
research
11/16/2022

CL2R: Compatible Lifelong Learning Representations

In this paper, we propose a method to partially mimic natural intelligen...
research
09/07/2018

HC-Net: Memory-based Incremental Dual-Network System for Continual learning

Training a neural network for a classification task typically assumes th...

Please sign up or login with your details

Forgot password? Click here to reset