Jointly Learning Truth-Conditional Denotations and Groundings using Parallel Attention

04/14/2021
by   Leon Bergen, et al.
0

We present a model that jointly learns the denotations of words together with their groundings using a truth-conditional semantics. Our model builds on the neurosymbolic approach of Mao et al. (2019), learning to ground objects in the CLEVR dataset (Johnson et al., 2017) using a novel parallel attention mechanism. The model achieves state of the art performance on visual question answering, learning to detect and ground objects with question performance as the only training signal. We also show that the model is able to learn flexible non-canonical groundings just by adjusting answers to questions in the training set.

READ FULL TEXT

page 3

page 6

page 7

page 8

page 13

page 14

page 15

research
04/16/2020

Bridging Anaphora Resolution as Question Answering

Most previous studies on bridging anaphora resolution (Poesio et al., 20...
research
10/18/2019

Relational Graph Representation Learning for Open-Domain Question Answering

We introduce a relational graph neural network with bi-directional atten...
research
06/01/2020

Probing Emergent Semantics in Predictive Agents via Question Answering

Recent work has shown how predictive modeling can endow agents with rich...
research
08/25/2018

Painting Outside the Box: Image Outpainting with GANs

The challenging task of image outpainting (extrapolation) has received c...
research
10/29/2019

Ordered Memory

Stack-augmented recurrent neural networks (RNNs) have been of interest t...
research
07/17/2023

PAT: Parallel Attention Transformer for Visual Question Answering in Vietnamese

We present in this paper a novel scheme for multimodal learning named th...
research
04/15/2017

Neural Paraphrase Identification of Questions with Noisy Pretraining

We present a solution to the problem of paraphrase identification of que...

Please sign up or login with your details

Forgot password? Click here to reset