Improving Visually Grounded Sentence Representations with Self-Attention

12/02/2017
by   Kang Min Yoo, et al.
0

Sentence representation models trained only on language could potentially suffer from the grounding problem. Recent work has shown promising results in improving the qualities of sentence representations by jointly training them with associated image features. However, the grounding capability is limited due to distant connection between input sentences and image features by the design of the architecture. In order to further close the gap, we propose applying self-attention mechanism to the sentence encoder to deepen the grounding effect. Our results on transfer tasks show that self-attentive encoders are better for visual grounding, as they exploit specific words with strong visual associations.

READ FULL TEXT
research
07/19/2017

Learning Visually Grounded Sentence Representations

We introduce a variety of models, trained on a supervised image captioni...
research
03/09/2017

A Structured Self-attentive Sentence Embedding

This paper proposes a new model for extracting an interpretable sentence...
research
06/26/2023

Learning with Difference Attention for Visually Grounded Self-supervised Representations

Recent works in self-supervised learning have shown impressive results o...
research
11/23/2017

Self-view Grounding Given a Narrated 360° Video

Narrated 360 videos are typically provided in many touring scenarios to ...
research
02/07/2020

Incorporating Visual Semantics into Sentence Representations within a Grounded Space

Language grounding is an active field aiming at enriching textual repres...
research
06/01/2019

Learning to Generate Grounded Image Captions without Localization Supervision

When generating a sentence description for an image, it frequently remai...
research
12/31/2019

Deep Attentive Ranking Networks for Learning to Order Sentences

We present an attention-based ranking framework for learning to order se...

Please sign up or login with your details

Forgot password? Click here to reset