Synthesizing Novel Pairs of Image and Text

12/18/2017
by   Jason Xie, et al.
0

Generating novel pairs of image and text is a problem that combines computer vision and natural language processing. In this paper, we present strategies for generating novel image and caption pairs based on existing captioning datasets. The model takes advantage of recent advances in generative adversarial networks and sequence-to-sequence modeling. We make generalizations to generate paired samples from multiple domains. Furthermore, we study cycles -- generating from image to text then back to image and vise versa, as well as its connection with autoencoders.

READ FULL TEXT
research
04/11/2018

Text2Colors: Guiding Image Colorization through Text-Driven Palette Generation

In this paper, we propose a novel approach to generate multiple color pa...
research
09/21/2016

Show and Tell: Lessons learned from the 2015 MSCOCO Image Captioning Challenge

Automatically describing the content of an image is a fundamental proble...
research
03/20/2017

I2T2I: Learning Text to Image Synthesis with Textual Data Augmentation

Translating information between text and image is a fundamental problem ...
research
05/31/2017

Adversarial Generation of Natural Language

Generative Adversarial Networks (GANs) have gathered a lot of attention ...
research
11/17/2014

Show and Tell: A Neural Image Caption Generator

Automatically describing the content of an image is a fundamental proble...
research
08/08/2023

The Five-Dollar Model: Generating Game Maps and Sprites from Sentence Embeddings

The five-dollar model is a lightweight text-to-image generative architec...
research
01/20/2022

A Computational Model for Machine Thinking

A machine thinking model is proposed in this report based on recent adva...

Please sign up or login with your details

Forgot password? Click here to reset