Variational Context: Exploiting Visual and Textual Context for Grounding Referring Expressions

07/08/2019
by   Yulei Niu, et al.
1

We focus on grounding (i.e., localizing or linking) referring expressions in images, e.g., "largest elephant standing behind baby elephant". This is a general yet challenging vision-language task since it does not only require the localization of objects, but also the multimodal comprehension of context -- visual attributes (e.g., "largest", "baby") and relationships (e.g., "behind") that help to distinguish the referent from other objects, especially those of the same category. Due to the exponential complexity involved in modeling the context associated with multiple image regions, existing work oversimplifies this task to pairwise region modeling by multiple instance learning. In this paper, we propose a variational Bayesian method, called Variational Context, to solve the problem of complex context modeling in referring expression grounding. Specifically, our framework exploits the reciprocal relation between the referent and context, i.e., either of them influences estimation of the posterior distribution of the other, and thereby the search space of context can be greatly reduced. In addition to reciprocity, our framework considers the semantic information of context, i.e., the referring expression can be reproduced based on the estimated context. We also extend the model to unsupervised setting where no annotation for the referent is available. Extensive experiments on various benchmarks show consistent improvement over state-of-the-art methods in both supervised and unsupervised settings.

READ FULL TEXT

page 1

page 5

page 7

page 8

page 9

page 10

page 12

page 13

research
12/05/2017

Grounding Referring Expressions in Images by Variational Context

We focus on grounding (i.e., localizing or linking) referring expression...
research
04/19/2020

Relationship-Embedded Representation Learning for Grounding Referring Expressions

Grounding referring expressions in images aims to locate the object inst...
research
06/11/2019

Cross-Modal Relationship Inference for Grounding Referring Expressions

Grounding referring expressions is a fundamental yet challenging task fa...
research
08/01/2016

Modeling Context Between Objects for Referring Expression Understanding

Referring expressions usually describe an object using properties of the...
research
02/17/2023

CK-Transformer: Commonsense Knowledge Enhanced Transformers for Referring Expression Comprehension

The task of multimodal referring expression comprehension (REC), aiming ...
research
06/09/2019

Referring Expression Grounding by Marginalizing Scene Graph Likelihood

We focus on the task of grounding referring expressions in images, e.g.,...
research
06/06/2023

Language Adaptive Weight Generation for Multi-task Visual Grounding

Although the impressive performance in visual grounding, the prevailing ...

Please sign up or login with your details

Forgot password? Click here to reset