RATT: Recurrent Attention to Transient Tasks for Continual Image Captioning

07/13/2020
by   Riccardo Del Chiaro, et al.
0

Research on continual learning has led to a variety of approaches to mitigating catastrophic forgetting in feed-forward classification networks. Until now surprisingly little attention has been focused on continual learning of recurrent models applied to problems like image captioning. In this paper we take a systematic look at continual learning of LSTM-based models for image captioning. We propose an attention-based approach that explicitly accommodates the transient nature of vocabularies in continual image captioning tasks – i.e. that task vocabularies are not disjoint. We call our method Recurrent Attention to Transient Tasks (RATT), and also show how to adapt continual learning approaches based on weight egularization and knowledge distillation to recurrent continual learning problems. We apply our approaches to incremental image captioning problem on two new continual learning benchmarks we define using the MS-COCO and Flickr30 datasets. Our results demonstrate that RATT is able to sequentially learn five captioning tasks while incurring no forgetting of previously learned ones.

READ FULL TEXT
research
09/19/2019

ContCap: A comprehensive framework for continual image captioning

While cutting-edge image captioning systems are increasingly describing ...
research
05/17/2021

Continual Learning with Echo State Networks

Continual Learning (CL) refers to a learning setup where data is non sta...
research
07/31/2023

Subspace Distillation for Continual Learning

An ultimate objective in continual learning is to preserve knowledge lea...
research
07/16/2021

Continual Learning for Automated Audio Captioning Using The Learning Without Forgetting Approach

Automated audio captioning (AAC) is the task of automatically creating t...
research
06/06/2023

Towards Adaptable and Interactive Image Captioning with Data Augmentation and Episodic Memory

Interactive machine learning (IML) is a beneficial learning paradigm in ...
research
09/26/2018

Batch-normalized Recurrent Highway Networks

Gradient control plays an important role in feed-forward networks applie...
research
06/15/2021

Natural continual learning: success is a journey, not (just) a destination

Biological agents are known to learn many different tasks over the cours...

Please sign up or login with your details

Forgot password? Click here to reset