TAG: Task-based Accumulated Gradients for Lifelong learning

05/11/2021
by   Pranshu Malviya, et al.
0

When an agent encounters a continual stream of new tasks in the lifelong learning setting, it leverages the knowledge it gained from the earlier tasks to help learn the new tasks better. In such a scenario, identifying an efficient knowledge representation becomes a challenging problem. Most research works propose to either store a subset of examples from the past tasks in a replay buffer, dedicate a separate set of parameters to each task or penalize excessive updates over parameters by introducing a regularization term. While existing methods employ the general task-agnostic stochastic gradient descent update rule, we propose a task-aware optimizer that adapts the learning rate based on the relatedness among tasks. We utilize the directions taken by the parameters during the updates by accumulating the gradients specific to each task. These task-based accumulated gradients act as a knowledge base that is maintained and updated throughout the stream. We empirically show that our proposed adaptive learning rate not only accounts for catastrophic forgetting but also allows positive backward transfer. We also show that our method performs better than several state-of-the-art methods in lifelong learning on complex datasets with a large number of tasks.

READ FULL TEXT

page 7

page 9

research
07/10/2022

Scaling the Number of Tasks in Continual Learning

Standard gradient descent algorithms applied to sequences of tasks are k...
research
11/28/2018

Experience Replay for Continual Learning

Continual learning is the problem of learning new tasks or knowledge whi...
research
05/17/2021

Layerwise Optimization by Gradient Decomposition for Continual Learning

Deep neural networks achieve state-of-the-art and sometimes super-human ...
research
11/28/2022

AdaTask: A Task-aware Adaptive Learning Rate Approach to Multi-task Learning

Multi-task learning (MTL) models have demonstrated impressive results in...
research
05/26/2022

Continual evaluation for lifelong learning: Identifying the stability gap

Introducing a time dependency on the data generating distribution has pr...
research
10/12/2022

Improving information retention in large scale online continual learning

Given a stream of data sampled from non-stationary distributions, online...
research
08/25/2021

Auxiliary Task Update Decomposition: The Good, The Bad and The Neutral

While deep learning has been very beneficial in data-rich settings, task...

Please sign up or login with your details

Forgot password? Click here to reset