Playable Environments: Video Manipulation in Space and Time

03/03/2022
by   Willi Menapace, et al.
2

We present Playable Environments - a new representation for interactive video generation and manipulation in space and time. With a single image at inference time, our novel framework allows the user to move objects in 3D while generating a video by providing a sequence of desired actions. The actions are learnt in an unsupervised manner. The camera can be controlled to get the desired viewpoint. Our method builds an environment state for each frame, which can be manipulated by our proposed action module and decoded back to the image space with volumetric rendering. To support diverse appearances of objects, we extend neural radiance fields with style-based modulation. Our method trains on a collection of various monocular videos requiring only the estimated camera parameters and 2D object locations. To set a challenging benchmark, we introduce two large scale video datasets with significant camera movements. As evidenced by our experiments, playable environments enable several creative applications not attainable by prior video synthesis works, including playable 3D video generation, stylization and manipulation. Further details, code and examples are available at https://willi-menapace.github.io/playable-environments-website

READ FULL TEXT

page 1

page 3

page 7

page 8

research
01/11/2022

HumanNeRF: Free-viewpoint Rendering of Moving People from Monocular Video

We introduce a free-viewpoint rendering method – HumanNeRF – that works ...
research
04/24/2023

HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video

We introduce HOSNeRF, a novel 360 free-viewpoint rendering method that r...
research
01/28/2021

Playable Video Generation

This paper introduces the unsupervised learning problem of playable vide...
research
09/17/2021

Diverse Generation from a Single Video Made Possible

Most advanced video generation and manipulation methods train on a large...
research
05/11/2022

Diverse Video Generation from a Single Video

GANs are able to perform generation and manipulation tasks, trained on a...
research
04/14/2022

Manually Acquiring Targets from Multiple Viewpoints Using Video Feedback

Objective: The effect of camera viewpoint was studied when performing vi...
research
07/07/2023

AutoDecoding Latent 3D Diffusion Models

We present a novel approach to the generation of static and articulated ...

Please sign up or login with your details

Forgot password? Click here to reset