Generating Images from Captions with Attention

11/09/2015
by   Elman Mansimov, et al.
0

Motivated by the recent progress in generative models, we introduce a model that generates images from natural language descriptions. The proposed model iteratively draws patches on a canvas, while attending to the relevant words in the description. After training on Microsoft COCO, we compare our model with several baseline generative models on image generation and retrieval tasks. We demonstrate that our model produces higher quality samples than other approaches and generates images with novel scene compositions corresponding to previously unseen captions in the dataset.

READ FULL TEXT

page 2

page 5

page 6

page 7

page 10

page 12

research
10/05/2018

CanvasGAN: A simple baseline for text to image generation by incrementally patching a canvas

We propose a new recurrent generative model for generating images from t...
research
04/26/2021

CAGAN: Text-To-Image Generation with Combined Attention GANs

Generating images according to natural language descriptions is a challe...
research
02/07/2022

Inference of captions from histopathological patches

Computational histopathology has made significant strides in the past fe...
research
05/14/2018

Normal Similarity Network for Generative Modelling

Gaussian distributions are commonly used as a key building block in many...
research
11/24/2021

MixSyn: Learning Composition and Style for Multi-Source Image Synthesis

Synthetic images created by generative models increase in quality and ex...
research
10/27/2022

How well can Text-to-Image Generative Models understand Ethical Natural Language Interventions?

Text-to-image generative models have achieved unprecedented success in g...
research
09/27/2022

What Does DALL-E 2 Know About Radiology?

Generative models such as DALL-E 2 could represent a promising future to...

Please sign up or login with your details

Forgot password? Click here to reset