Towards the Probabilistic Fusion of Learned Priors into Standard Pipelines for 3D Reconstruction

07/27/2022
by   Tristan Laidlow, et al.
3

The best way to combine the results of deep learning with standard 3D reconstruction pipelines remains an open problem. While systems that pass the output of traditional multi-view stereo approaches to a network for regularisation or refinement currently seem to get the best results, it may be preferable to treat deep neural networks as separate components whose results can be probabilistically fused into geometry-based systems. Unfortunately, the error models required to do this type of fusion are not well understood, with many different approaches being put forward. Recently, a few systems have achieved good results by having their networks predict probability distributions rather than single values. We propose using this approach to fuse a learned single-view depth prior into a standard 3D reconstruction system. Our system is capable of incrementally producing dense depth maps for a set of keyframes. We train a deep neural network to predict discrete, nonparametric probability distributions for the depth of each pixel from a single image. We then fuse this "probability volume" with another probability volume based on the photometric consistency between subsequent frames and the keyframe image. We argue that combining the probability volumes from these two sources will result in a volume that is better conditioned. To extract depth maps from the volume, we minimise a cost function that includes a regularisation term based on network predicted surface normals and occlusion boundaries. Through a series of experiments, we demonstrate that each of these components improves the overall performance of the system.

READ FULL TEXT

page 1

page 4

page 6

research
08/19/2021

VolumeFusion: Deep Depth Fusion for 3D Scene Reconstruction

To reconstruct a 3D scene from a set of calibrated views, traditional mu...
research
11/09/2022

ReFu: Refine and Fuse the Unobserved View for Detail-Preserving Single-Image 3D Human Reconstruction

Single-image 3D human reconstruction aims to reconstruct the 3D textured...
research
01/31/2019

Pix2Vox: Context-aware 3D Reconstruction from Single and Multi-view Images

Recovering the 3D representation of an object from single-view or multi-...
research
06/16/2023

C2F2NeUS: Cascade Cost Frustum Fusion for High Fidelity and Generalizable Neural Surface Reconstruction

There is an emerging effort to combine the two popular technical paths, ...
research
12/26/2019

Learning Inverse Depth Regression for Multi-View Stereo with Correlation Cost Volume

Deep learning has shown to be effective for depth inference in multi-vie...
research
05/24/2023

Incremental Dense Reconstruction from Monocular Video with Guided Sparse Feature Volume Fusion

Incrementally recovering 3D dense structures from monocular videos is of...
research
03/15/2019

SceneCode: Monocular Dense Semantic Reconstruction using Learned Encoded Scene Representations

Systems which incrementally create 3D semantic maps from image sequences...

Please sign up or login with your details

Forgot password? Click here to reset