CanvasGAN: A simple baseline for text to image generation by incrementally patching a canvas

10/05/2018
by   Amanpreet Singh, et al.
0

We propose a new recurrent generative model for generating images from text captions while attending on specific parts of text captions. Our model creates images by incrementally adding patches on a "canvas" while attending on words from text caption at each timestep. Finally, the canvas is passed through an upscaling network to generate images. We also introduce a new method for generating visual-semantic sentence embeddings based on self-attention over text. We compare our model's generated images with those generated Reed et. al.'s model and show that our model is a stronger baseline for text to image generation tasks.

READ FULL TEXT

page 9

page 14

research
11/09/2015

Generating Images from Captions with Attention

Motivated by the recent progress in generative models, we introduce a mo...
research
06/05/2023

Composition and Deformance: Measuring Imageability with a Text-to-Image Model

Although psycholinguists and psychologists have long studied the tendenc...
research
02/16/2016

Generating images with recurrent adversarial networks

Gatys et al. (2015) showed that optimizing pixels to match features in a...
research
08/08/2023

The Five-Dollar Model: Generating Game Maps and Sprites from Sentence Embeddings

The five-dollar model is a lightweight text-to-image generative architec...
research
02/22/2018

ChatPainter: Improving Text to Image Generation using Dialogue

Synthesizing realistic images from text descriptions on a dataset like M...
research
02/07/2022

Inference of captions from histopathological patches

Computational histopathology has made significant strides in the past fe...
research
05/09/2019

Interactive Image Generation Using Scene Graphs

Recent years have witnessed some exciting developments in the domain of ...

Please sign up or login with your details

Forgot password? Click here to reset