Eliminating Catastrophic Interference with Biased Competition

07/03/2020
by   Amelia Elizabeth Pollard, et al.
0

We present here a model to take advantage of the multi-task nature of complex datasets by learning to separate tasks and subtasks in and end to end manner by biasing competitive interactions in the network. This method does not require additional labelling or reformatting of data in a dataset. We propose an alternate view to the monolithic one-task-fits-all learning of multi-task problems, and describe a model based on a theory of neuronal attention from neuroscience, proposed by Desimone. We create and exhibit a new toy dataset, based on the MNIST dataset, which we call MNIST-QA, for testing Visual Question Answering architectures in a low-dimensional environment while preserving the more difficult components of the Visual Question Answering task, and demonstrate the proposed network architecture on this new dataset, as well as on COCO-QA and DAQUAR-FULL. We then demonstrate that this model eliminates catastrophic interference between tasks on a newly created toy dataset and provides competitive results in the Visual Question Answering space. We provide further evidence that Visual Question Answering can be approached as a multi-task problem, and demonstrate that this new architecture based on the Biased Competition model is capable of learning to separate and learn the tasks in an end-to-end fashion without the need for task labels.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/03/2020

Visual Question Answering as a Multi-Task Problem

Visual Question Answering(VQA) is a highly complex problem set, relying ...
research
04/27/2019

Using Context Information to Enhance Simple Question Answering

With the rapid development of knowledge bases(KBs),question answering(QA...
research
09/05/2018

TVQA: Localized, Compositional Video Question Answering

Recent years have witnessed an increasing interest in image-based questi...
research
08/04/2017

MemexQA: Visual Memex Question Answering

This paper proposes a new task, MemexQA: given a collection of photos or...
research
04/20/2021

Efficient Retrieval Optimized Multi-task Learning

Recently, there have been significant advances in neural methods for tac...
research
06/22/2021

Fine-tune the Entire RAG Architecture (including DPR retriever) for Question-Answering

In this paper, we illustrate how to fine-tune the entire Retrieval Augme...
research
04/13/2021

Structural analysis of an all-purpose question answering model

Attention is a key component of the now ubiquitous pre-trained language ...

Please sign up or login with your details

Forgot password? Click here to reset