MixerGAN: An MLP-Based Architecture for Unpaired Image-to-Image Translation

05/28/2021
by   George Cazenavette, et al.
0

While attention-based transformer networks achieve unparalleled success in nearly all language tasks, the large number of tokens coupled with the quadratic activation memory usage makes them prohibitive for visual tasks. As such, while language-to-language translation has been revolutionized by the transformer model, convolutional networks remain the de facto solution for image-to-image translation. The recently proposed MLP-Mixer architecture alleviates some of the speed and memory issues associated with attention-based networks while still retaining the long-range connections that make transformer models desirable. Leveraging this efficient alternative to self-attention, we propose a new unpaired image-to-image translation model called MixerGAN: a simpler MLP-based architecture that considers long-distance relationships between pixels without the need for expensive attention mechanisms. Quantitative and qualitative analysis shows that MixerGAN achieves competitive results when compared to prior convolutional-based methods.

READ FULL TEXT

page 1

page 3

page 8

page 9

research
01/24/2019

Unsupervised Image-to-Image Translation with Self-Attention Networks

Unsupervised image translation aims to learn the transformation from a s...
research
03/30/2022

ITTR: Unpaired Image-to-Image Translation with Transformers

Unpaired image-to-image translation is to translate an image from a sour...
research
02/07/2019

Reversible GANs for Memory-efficient Image-to-Image Translation

The Pix2pix and CycleGAN losses have vastly improved the qualitative and...
research
11/10/2021

Palette: Image-to-Image Diffusion Models

We introduce Palette, a simple and general framework for image-to-image ...
research
03/30/2022

InstaFormer: Instance-Aware Image-to-Image Translation with Transformer

We present a novel Transformer-based network architecture for instance-a...
research
07/25/2019

U-GAT-IT: Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization for Image-to-Image Translation

We propose a novel method for unsupervised image-to-image translation, w...
research
07/09/2023

Latent Graph Attention for Enhanced Spatial Context

Global contexts in images are quite valuable in image-to-image translati...

Please sign up or login with your details

Forgot password? Click here to reset