DreamSparse: Escaping from Plato's Cave with 2D Frozen Diffusion Model Given Sparse Views

06/06/2023
by   Paul Yoo, et al.
0

Synthesizing novel view images from a few views is a challenging but practical problem. Existing methods often struggle with producing high-quality results or necessitate per-object optimization in such few-view settings due to the insufficient information provided. In this work, we explore leveraging the strong 2D priors in pre-trained diffusion models for synthesizing novel view images. 2D diffusion models, nevertheless, lack 3D awareness, leading to distorted image synthesis and compromising the identity. To address these problems, we propose DreamSparse, a framework that enables the frozen pre-trained diffusion model to generate geometry and identity-consistent novel view image. Specifically, DreamSparse incorporates a geometry module designed to capture 3D features from sparse views as a 3D prior. Subsequently, a spatial guidance model is introduced to convert these 3D feature maps into spatial information for the generative process. This information is then used to guide the pre-trained diffusion model, enabling it to generate geometrically consistent images without tuning it. Leveraging the strong image priors in the pre-trained diffusion models, DreamSparse is capable of synthesizing high-quality novel views for both object and scene-level images and generalising to open-set images. Experimental results demonstrate that our framework can effectively synthesize novel view images from sparse views and outperforms baselines in both trained and open-set category images. More results can be found on our project page: https://sites.google.com/view/dreamsparse-webpage.

READ FULL TEXT

page 1

page 4

page 6

page 8

page 9

research
08/27/2023

Sparse3D: Distilling Multiview-Consistent Diffusion for Object Reconstruction from Sparse Views

Reconstructing 3D objects from extremely sparse views is a long-standing...
research
06/10/2022

NeRF-In: Free-Form NeRF Inpainting with RGB-D Priors

Though Neural Radiance Field (NeRF) demonstrates compelling novel view s...
research
12/12/2022

Generative Scene Synthesis via Incremental View Inpainting using RGBD Diffusion Models

We address the challenge of recovering an underlying scene geometry and ...
research
11/29/2022

NeuralLift-360: Lifting An In-the-wild 2D Photo to A 3D Object with 360° Views

Virtual reality and augmented reality (XR) bring increasing demand for 3...
research
05/30/2023

HiFA: High-fidelity Text-to-3D with Advanced Diffusion Guidance

Automatic text-to-3D synthesis has achieved remarkable advancements thro...
research
07/11/2023

My3DGen: Building Lightweight Personalized 3D Generative Model

Our paper presents My3DGen, a practical system for creating a personaliz...
research
04/08/2023

RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors

Existing dehazing approaches struggle to process real-world hazy images ...

Please sign up or login with your details

Forgot password? Click here to reset