ZM-Net: Real-time Zero-shot Image Manipulation Network

03/21/2017
by   Hao Wang, et al.
0

Many problems in image processing and computer vision (e.g. colorization, style transfer) can be posed as 'manipulating' an input image into a corresponding output image given a user-specified guiding signal. A holy-grail solution towards generic image manipulation should be able to efficiently alter an input image with any personalized signals (even signals unseen during training), such as diverse paintings and arbitrary descriptive attributes. However, existing methods are either inefficient to simultaneously process multiple signals (let alone generalize to unseen signals), or unable to handle signals from other modalities. In this paper, we make the first attempt to address the zero-shot image manipulation task. We cast this problem as manipulating an input image according to a parametric model whose key parameters can be conditionally generated from any guiding signal (even unseen ones). To this end, we propose the Zero-shot Manipulation Net (ZM-Net), a fully-differentiable architecture that jointly optimizes an image-transformation network (TNet) and a parameter network (PNet). The PNet learns to generate key transformation parameters for the TNet given any guiding signal while the TNet performs fast zero-shot image manipulation according to both signal-dependent parameters from the PNet and signal-invariant parameters from the TNet itself. Extensive experiments show that our ZM-Net can perform high-quality image manipulation conditioned on different forms of guiding signals (e.g. style images and attributes) in real-time (tens of milliseconds per image) even for unseen signals. Moreover, a large-scale style dataset with over 20,000 style images is also constructed to promote further research.

READ FULL TEXT

page 1

page 5

page 7

page 8

research
05/10/2018

Avatar-Net: Multi-scale Zero-shot Style Transfer by Feature Decoration

Zero-shot artistic style transfer is an important image synthesis proble...
research
03/15/2023

Zero-Shot Contrastive Loss for Text-Guided Diffusion Image Style Transfer

Diffusion models have shown great promise in text-guided image style tra...
research
03/19/2023

StyleRF: Zero-shot 3D Style Transfer of Neural Radiance Fields

3D style transfer aims to render stylized novel views of a 3D scene with...
research
03/16/2018

Deep Multiple Instance Learning for Zero-shot Image Tagging

In-line with the success of deep learning on traditional recognition pro...
research
08/17/2019

Zero Shot Learning for Multi-Modal Real Time Image Registration

In this report we present an unsupervised image registration framework, ...
research
08/22/2018

Manipulating Attributes of Natural Scenes via Hallucination

In this study, we explore building a two-stage framework for enabling us...

Please sign up or login with your details

Forgot password? Click here to reset