RenderDiffusion: Image Diffusion for 3D Reconstruction, Inpainting and Generation

11/17/2022
by   Titas Anciukevicius, et al.
1

Diffusion models currently achieve state-of-the-art performance for both conditional and unconditional image generation. However, so far, image diffusion models do not support tasks required for 3D understanding, such as view-consistent 3D generation or single-view object reconstruction. In this paper, we present RenderDiffusion as the first diffusion model for 3D generation and inference that can be trained using only monocular 2D supervision. At the heart of our method is a novel image denoising architecture that generates and renders an intermediate three-dimensional representation of a scene in each denoising step. This enforces a strong inductive structure into the diffusion process that gives us a 3D consistent representation while only requiring 2D supervision. The resulting 3D representation can be rendered from any viewpoint. We evaluate RenderDiffusion on ShapeNet and Clevr datasets and show competitive performance for generation of 3D scenes and inference of 3D scenes from 2D images. Additionally, our diffusion-based approach allows us to use 2D inpainting to edit 3D scenes. We believe that our work promises to enable full 3D generation at scale when trained on massive image collections, thus circumventing the need to have large-scale 3D model collections for supervision.

READ FULL TEXT

page 1

page 5

page 6

page 7

page 8

research
09/18/2023

Gradpaint: Gradient-Guided Inpainting with Diffusion Models

Denoising Diffusion Probabilistic Models (DDPMs) have recently achieved ...
research
11/29/2022

SinDDM: A Single Image Denoising Diffusion Model

Denoising diffusion models (DDMs) have led to staggering performance lea...
research
03/23/2020

Pix2Shape – Towards Unsupervised Learning of 3D Scenes from Images using a View-based Representation

We infer and generate three-dimensional (3D) scene information from a si...
research
12/01/2022

SparseFusion: Distilling View-conditioned Diffusion for 3D Reconstruction

We propose SparseFusion, a sparse view 3D reconstruction approach that u...
research
06/08/2023

ADDP: Learning General Representations for Image Recognition and Generation with Alternating Denoising Diffusion Process

Image recognition and generation have long been developed independently ...
research
07/25/2023

Fake It Without Making It: Conditioned Face Generation for Accurate 3D Face Shape Estimation

Accurate 3D face shape estimation is an enabling technology with applica...
research
06/16/2023

Drag-guided diffusion models for vehicle image generation

Denoising diffusion models trained at web-scale have revolutionized imag...

Please sign up or login with your details

Forgot password? Click here to reset