GenLayNeRF: Generalizable Layered Representations with 3D Model Alignment for Multi-Human View Synthesis

09/20/2023
by   Youssef Abdelkareem, et al.
0

Novel view synthesis (NVS) of multi-human scenes imposes challenges due to the complex inter-human occlusions. Layered representations handle the complexities by dividing the scene into multi-layered radiance fields, however, they are mainly constrained to per-scene optimization making them inefficient. Generalizable human view synthesis methods combine the pre-fitted 3D human meshes with image features to reach generalization, yet they are mainly designed to operate on single-human scenes. Another drawback is the reliance on multi-step optimization techniques for parametric pre-fitting of the 3D body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views. In this work, we propose, GenLayNeRF, a generalizable layered scene representation for free-viewpoint rendering of multiple human subjects which requires no per-scene optimization and very sparse views as input. We divide the scene into multi-human layers anchored by the 3D body meshes. We then ensure pixel-level alignment of the body models with the input views through a novel end-to-end trainable module that carries out iterative parametric correction coupled with multi-view feature fusion to produce aligned 3D models. For NVS, we extract point-wise image-aligned and human-anchored features which are correlated and fused using self-attention and cross-attention modules. We augment low-level RGB values into the features with an attention-based RGB fusion module. To evaluate our approach, we construct two multi-human view synthesis datasets; DeepMultiSyn and ZJU-MultiHuman. The results indicate that our proposed approach outperforms generalizable and non-human per-scene NeRF methods while performing at par with layered per-scene methods without test time optimization.

READ FULL TEXT

page 3

page 7

page 8

research
03/03/2023

Multi-Plane Neural Radiance Fields for Novel View Synthesis

Novel view synthesis is a long-standing problem that revolves around ren...
research
09/15/2021

Neural Human Performer: Learning Generalizable Radiance Fields for Human Performance Rendering

In this paper, we aim at synthesizing a free-viewpoint video of an arbit...
research
03/31/2023

Efficient View Synthesis and 3D-based Multi-Frame Denoising with Multiplane Feature Representations

While current multi-frame restoration methods combine information from m...
research
10/04/2022

Self-improving Multiplane-to-layer Images for Novel View Synthesis

We present a new method for lightweight novel-view synthesis that genera...
research
09/10/2023

SC-NeRF: Self-Correcting Neural Radiance Field with Sparse Views

In recent studies, the generalization of neural radiance fields for nove...
research
03/23/2023

Semantic Ray: Learning a Generalizable Semantic Field with Cross-Reprojection Attention

In this paper, we aim to learn a semantic radiance field from multiple s...
research
08/11/2020

VI-Net: View-Invariant Quality of Human Movement Assessment

We propose a view-invariant method towards the assessment of the quality...

Please sign up or login with your details

Forgot password? Click here to reset