RefineCap: Concept-Aware Refinement for Image Captioning

09/08/2021
by   Yekun Chai, et al.
0

Automatically translating images to texts involves image scene understanding and language modeling. In this paper, we propose a novel model, termed RefineCap, that refines the output vocabulary of the language decoder using decoder-guided visual semantics, and implicitly learns the mapping between visual tag words and images. The proposed Visual-Concept Refinement method can allow the generator to attend to semantic details in the image, thereby generating more semantically descriptive captions. Our model achieves superior performance on the MS-COCO dataset in comparison with previous visual-concept based models.

READ FULL TEXT

page 4

page 8

page 9

research
08/30/2019

Reflective Decoding Network for Image Captioning

State-of-the-art image captioning methods mostly focus on improving visu...
research
08/04/2021

Question-controlled Text-aware Image Captioning

For an image with multiple scene texts, different people may be interest...
research
11/21/2015

Mapping Images to Sentiment Adjective Noun Pairs with Factorized Neural Nets

We consider the visual sentiment task of mapping an image to an adjectiv...
research
02/03/2021

L2C: Describing Visual Differences Needs Semantic Understanding of Individuals

Recent advances in language and vision push forward the research of capt...
research
09/29/2021

Geometry-Entangled Visual Semantic Transformer for Image Captioning

Recent advancements of image captioning have featured Visual-Semantic Fu...
research
10/02/2020

CAPTION: Correction by Analyses, POS-Tagging and Interpretation of Objects using only Nouns

Recently, Deep Learning (DL) methods have shown an excellent performance...
research
03/06/2019

Image captioning with weakly-supervised attention penalty

Stories are essential for genealogy research since they can help build e...

Please sign up or login with your details

Forgot password? Click here to reset