Conditional Generative ConvNets for Exemplar-based Texture Synthesis

12/17/2019
by   Zi-Ming Wang, et al.
17

The goal of exemplar-based texture synthesis is to generate texture images that are visually similar to a given exemplar. Recently, promising results have been reported by methods relying on convolutional neural networks (ConvNets) pretrained on large-scale image datasets. However, these methods have difficulties in synthesizing image textures with non-local structures and extending to dynamic or sound textures. In this paper, we present a conditional generative ConvNet (cgCNN) model which combines deep statistics and the probabilistic framework of generative ConvNet (gCNN) model. Given a texture exemplar, the cgCNN model defines a conditional distribution using deep statistics of a ConvNet, and synthesize new textures by sampling from the conditional distribution. In contrast to previous deep texture models, the proposed cgCNN dose not rely on pre-trained ConvNets but learns the weights of ConvNets for each input exemplar instead. As a result, the cgCNN model can synthesize high quality dynamic, sound and image textures in a unified manner. We also explore the theoretical connections between our model and other texture models. Further investigations show that the cgCNN model can be easily generalized to texture expansion and inpainting. Extensive experiments demonstrate that our model can achieve better or at least comparable results than the state-of-the-art methods.

READ FULL TEXT

page 1

page 7

page 8

page 9

page 10

page 11

page 12

page 13

research
05/17/2020

Co-occurrence Based Texture Synthesis

We model local texture patterns using the co-occurrence statistics of pi...
research
06/30/2020

Deep Geometric Texture Synthesis

Recently, deep generative adversarial networks for image generation have...
research
02/03/2018

Learning the Synthesizability of Dynamic Texture Samples

A dynamic texture (DT) refers to a sequence of images that exhibit tempo...
research
10/08/2015

Texture Modelling with Nested High-order Markov-Gibbs Random Fields

Currently, Markov-Gibbs random field (MGRF) image models which include h...
research
03/05/2017

Diversified Texture Synthesis with Feed-forward Networks

Recent progresses on deep discriminative and generative modeling have sh...
research
07/29/2018

Texture Mixing by Interpolating Deep Statistics via Gaussian Models

Recently, enthusiastic studies have devoted to texture synthesis using d...
research
06/21/2017

Two-Stream Convolutional Networks for Dynamic Texture Synthesis

We introduce a two-stream model for dynamic texture synthesis. Our model...

Please sign up or login with your details

Forgot password? Click here to reset