DynaST: Dynamic Sparse Transformer for Exemplar-Guided Image Generation

07/13/2022
by   Songhua Liu, et al.
0

One key challenge of exemplar-guided image generation lies in establishing fine-grained correspondences between input and guided images. Prior approaches, despite the promising results, have relied on either estimating dense attention to compute per-point matching, which is limited to only coarse scales due to the quadratic memory cost, or fixing the number of correspondences to achieve linear complexity, which lacks flexibility. In this paper, we propose a dynamic sparse attention based Transformer model, termed Dynamic Sparse Transformer (DynaST), to achieve fine-level matching with favorable efficiency. The heart of our approach is a novel dynamic-attention unit, dedicated to covering the variation on the optimal number of tokens one position should focus on. Specifically, DynaST leverages the multi-layer nature of Transformer structure, and performs the dynamic attention scheme in a cascaded manner to refine matching results and synthesize visually-pleasing outputs. In addition, we introduce a unified training objective for DynaST, making it a versatile reference-based image translation framework for both supervised and unsupervised scenarios. Extensive experiments on three applications, pose-guided person image generation, edge-based face synthesis, and undistorted image style transfer, demonstrate that DynaST achieves superior performance in local details, outperforming the state of the art while reducing the computational cost significantly. Our code is available at https://github.com/Huage001/DynaST

READ FULL TEXT

page 2

page 10

page 13

page 21

page 22

page 23

page 24

page 25

research
03/30/2023

Masked and Adaptive Transformer for Exemplar Based Image Translation

We present a novel framework for exemplar based image translation. Recen...
research
07/07/2021

Bi-level Feature Alignment for Versatile Image Translation and Manipulation

Generative adversarial networks (GANs) have achieved great success in im...
research
08/27/2020

Deep Spatial Transformation for Pose-Guided Person Image Generation and Animation

Pose-guided person image generation and animation aim to transform a sou...
research
03/06/2022

Exploring Dual-task Correlation for Pose Guided Person Image Generation

Pose Guided Person Image Generation (PGPIG) is the task of transforming ...
research
12/20/2021

StyleSwin: Transformer-based GAN for High-resolution Image Generation

Despite the tantalizing success in a broad of vision tasks, transformers...
research
11/10/2022

StyleNAT: Giving Each Head a New Perspective

Image generation has been a long sought-after but challenging task, and ...
research
01/31/2021

Cascade Network with Guided Loss and Hybrid Attention for Finding Good Correspondences

Finding good correspondences is a critical prerequisite in many feature ...

Please sign up or login with your details

Forgot password? Click here to reset