MagicMix: Semantic Mixing with Diffusion Models

10/28/2022
by   Jun Hao Liew, et al.
0

Have you ever imagined what a corgi-alike coffee machine or a tiger-alike rabbit would look like? In this work, we attempt to answer these questions by exploring a new task called semantic mixing, aiming at blending two different semantics to create a new concept (e.g., corgi + coffee machine – > corgi-alike coffee machine). Unlike style transfer, where an image is stylized according to the reference style without changing the image content, semantic blending mixes two different concepts in a semantic manner to synthesize a novel concept while preserving the spatial layout and geometry. To this end, we present MagicMix, a simple yet effective solution based on pre-trained text-conditioned diffusion models. Motivated by the progressive generation property of diffusion models where layout/shape emerges at early denoising steps while semantically meaningful details appear at later steps during the denoising process, our method first obtains a coarse layout (either by corrupting an image or denoising from a pure Gaussian noise given a text prompt), followed by injection of conditional prompt for semantic mixing. Our method does not require any spatial mask or re-training, yet is able to synthesize novel objects with high fidelity. To improve the mixing quality, we further devise two simple strategies to provide better control and flexibility over the synthesized content. With our method, we present our results over diverse downstream applications, including semantic style transfer, novel object synthesis, breed mixing, and concept removal, demonstrating the flexibility of our method. More results can be found on the project page https://magicmix.github.io

READ FULL TEXT

page 1

page 7

page 9

page 10

page 11

page 12

page 13

page 14

research
02/28/2023

Towards Enhanced Controllability of Diffusion Models

Denoising Diffusion models have shown remarkable capabilities in generat...
research
01/12/2019

Automated Deep Photo Style Transfer

Photorealism is a complex concept that cannot easily be formulated mathe...
research
02/14/2023

DiffFashion: Reference-based Fashion Design with Structure-aware Transfer by Diffusion Models

Image-based fashion design with AI techniques has attracted increasing a...
research
05/09/2023

Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style Transfer

Large-scale text-to-video diffusion models have demonstrated an exceptio...
research
11/29/2022

SinDDM: A Single Image Denoising Diffusion Model

Denoising diffusion models (DDMs) have led to staggering performance lea...
research
01/28/2023

SEGA: Instructing Diffusion using Semantic Dimensions

Text-to-image diffusion models have recently received a lot of interest ...
research
04/16/2021

ScreenSeg: On-Device Screenshot Layout Analysis

We propose a novel end-to-end solution that performs a Hierarchical Layo...

Please sign up or login with your details

Forgot password? Click here to reset