Cycle Text-To-Image GAN with BERT

03/26/2020
by   Trevor Tsue, et al.
7

We explore novel approaches to the task of image generation from their respective captions, building on state-of-the-art GAN architectures. Particularly, we baseline our models with the Attention-based GANs that learn attention mappings from words to image features. To better capture the features of the descriptions, we then built a novel cyclic design that learns an inverse function to maps the image back to original caption. Additionally, we incorporated recently developed BERT pretrained word embeddings as our initial text featurizer and observe a noticeable improvement in qualitative and quantitative performance compared to the Attention GAN baseline.

READ FULL TEXT

page 2

page 6

research
08/03/2021

Cycle-Consistent Inverse GAN for Text-to-Image Synthesis

This paper investigates an open research task of text-to-image synthesis...
research
04/01/2021

Text to Image Generation with Semantic-Spatial Aware GAN

A text to image generation (T2I) model aims to generate photo-realistic ...
research
09/19/2021

ComicGAN: Text-to-Comic Generative Adversarial Network

Drawing and annotating comic illustrations is a complex and difficult pr...
research
06/01/2023

ViCo: Detail-Preserving Visual Condition for Personalized Text-to-Image Generation

Personalized text-to-image generation using diffusion models has recentl...
research
06/06/2019

Visualizing and Measuring the Geometry of BERT

Transformer architectures show significant promise for natural language ...
research
01/18/2020

Text-to-Image Generation with Attention Based Recurrent Neural Networks

Conditional image modeling based on textual descriptions is a relatively...
research
01/12/2021

Cross-Modal Contrastive Learning for Text-to-Image Generation

The output of text-to-image synthesis systems should be coherent, clear,...

Please sign up or login with your details

Forgot password? Click here to reset