Image Representations and New Domains in Neural Image Captioning

08/09/2015
by   Jack Hessel, et al.
0

We examine the possibility that recent promising results in automatic caption generation are due primarily to language models. By varying image representation quality produced by a convolutional neural network, we find that a state-of-the-art neural captioning algorithm is able to produce quality captions even when provided with surprisingly poor image representations. We replicate this result in a new, fine-grained, transfer learned captioning domain, consisting of 66K recipe image/title pairs. We also provide some experiments regarding the appropriateness of datasets for automatic captioning, and find that having multiple captions per image is beneficial, but not an absolute requirement.

READ FULL TEXT

page 4

page 5

research
05/29/2020

Controlling Length in Image Captioning

We develop and evaluate captioning models that allow control of caption ...
research
07/21/2018

What is not where: the challenge of integrating spatial representations into deep learning architectures

This paper examines to what degree current deep learning architectures f...
research
10/10/2022

Generating image captions with external encyclopedic knowledge

Accurately reporting what objects are depicted in an image is largely a ...
research
06/28/2023

VisText: A Benchmark for Semantically Rich Chart Captioning

Captions that describe or explain charts help improve recall and compreh...
research
05/11/2023

Simple Token-Level Confidence Improves Caption Correctness

The ability to judge whether a caption correctly describes an image is a...
research
09/27/2018

Vector Learning for Cross Domain Representations

Recently, generative adversarial networks have gained a lot of popularit...
research
08/12/2016

DeepDiary: Automatic Caption Generation for Lifelogging Image Streams

Lifelogging cameras capture everyday life from a first-person perspectiv...

Please sign up or login with your details

Forgot password? Click here to reset