NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors

12/06/2022
by   Congyue Deng, et al.
0

2D-to-3D reconstruction is an ill-posed problem, yet humans are good at solving this problem due to their prior knowledge of the 3D world developed over years. Driven by this observation, we propose NeRDi, a single-view NeRF synthesis framework with general image priors from 2D diffusion models. Formulating single-view reconstruction as an image-conditioned 3D generation problem, we optimize the NeRF representations by minimizing a diffusion loss on its arbitrary view renderings with a pretrained image diffusion model under the input-view constraint. We leverage off-the-shelf vision-language models and introduce a two-section language guidance as conditioning inputs to the diffusion model. This is essentially helpful for improving multiview content coherence as it narrows down the general image prior conditioned on the semantic and visual features of the single-view input image. Additionally, we introduce a geometric loss based on estimated depth maps to regularize the underlying 3D geometry of the NeRF. Experimental results on the DTU MVS dataset show that our method can synthesize novel views with higher quality even compared to existing methods trained on this dataset. We also demonstrate our generalizability in zero-shot NeRF synthesis for in-the-wild images.

READ FULL TEXT

page 1

page 5

page 6

page 7

page 8

page 10

research
03/20/2023

Zero-1-to-3: Zero-shot One Image to 3D Object

We introduce Zero-1-to-3, a framework for changing the camera viewpoint ...
research
09/14/2023

Viewpoint Textual Inversion: Unleashing Novel View Synthesis with Pretrained 2D Diffusion Models

Text-to-image diffusion models understand spatial relationship between o...
research
06/26/2023

Fuzzy-Conditioned Diffusion and Diffusion Projection Attention Applied to Facial Image Correction

Image diffusion has recently shown remarkable performance in image synth...
research
03/08/2017

Transformation-Grounded Image Generation Network for Novel 3D View Synthesis

We present a transformation-grounded image generation network for novel ...
research
10/06/2022

Novel View Synthesis with Diffusion Models

We present 3DiM, a diffusion model for 3D novel view synthesis, which is...
research
04/13/2023

Learning Controllable 3D Diffusion Models from Single-view Images

Diffusion models have recently become the de-facto approach for generati...
research
12/12/2018

Extreme View Synthesis

We present Extreme View Synthesis, a solution for novel view extrapolati...

Please sign up or login with your details

Forgot password? Click here to reset