Actively Seeking and Learning from Live Data

04/05/2019
by   Damien Teney, et al.
34

One of the key limitations of traditional machine learning methods is their requirement for training data that exemplifies all the information to be learned. This is a particular problem for visual question answering methods, which may be asked questions about virtually anything. The approach we propose is a step toward overcoming this limitation by searching for the information required at test time. The resulting method dynamically utilizes data from an external source, such as a large set of questions/answers or images/captions. Concretely, we learn a set of base weights for a simple VQA model, that are specifically adapted to a given question with the information specifically retrieved for this question. The adaptation process leverages recent advances in gradient-based meta learning and contributions for efficient retrieval and cross-domain adaptation. We surpass the state-of-the-art on the VQA-CP v2 benchmark and demonstrate our approach to be intrinsically more robust to out-of-distribution test data. We demonstrate the use of external non-VQA data using the MS COCO captioning dataset to support the answering process. This approach opens a new avenue for open-domain VQA systems that interface with diverse sources of data.

READ FULL TEXT

page 1

page 8

page 14

page 15

page 16

page 17

page 18

page 19

research
11/22/2017

Visual Question Answering as a Meta Learning Task

The predominant approach to Visual Question Answering (VQA) demands that...
research
12/04/2017

Learning by Asking Questions

We introduce an interactive learning framework for the development and t...
research
05/22/2018

Joint Image Captioning and Question Answering

Answering visual questions need acquire daily common knowledge and model...
research
05/19/2021

Multiple Meta-model Quantifying for Medical Visual Question Answering

Transfer learning is an important step to extract meaningful features an...
research
09/26/2019

Overcoming Data Limitation in Medical Visual Question Answering

Traditional approaches for Visual Question Answering (VQA) require large...
research
06/13/2023

AVIS: Autonomous Visual Information Seeking with Large Language Models

In this paper, we propose an autonomous information seeking visual quest...
research
03/29/2021

Domain-robust VQA with diverse datasets and methods but no target labels

The observation that computer vision methods overfit to dataset specific...

Please sign up or login with your details

Forgot password? Click here to reset