TeCH: Text-guided Reconstruction of Lifelike Clothed Humans

08/16/2023
by   Yangyi Huang, et al.
0

Despite recent research advancements in reconstructing clothed humans from a single image, accurately restoring the "unseen regions" with high-level details remains an unsolved challenge that lacks attention. Existing methods often generate overly smooth back-side surfaces with a blurry texture. But how to effectively capture all visual attributes of an individual from a single image, which are sufficient to reconstruct unseen areas (e.g., the back view)? Motivated by the power of foundation models, TeCH reconstructs the 3D human by leveraging 1) descriptive text prompts (e.g., garments, colors, hairstyles) which are automatically generated via a garment parsing model and Visual Question Answering (VQA), 2) a personalized fine-tuned Text-to-Image diffusion model (T2I) which learns the "indescribable" appearance. To represent high-resolution 3D clothed humans at an affordable cost, we propose a hybrid 3D representation based on DMTet, which consists of an explicit body shape grid and an implicit distance field. Guided by the descriptive prompts + personalized T2I diffusion model, the geometry and texture of the 3D humans are optimized through multi-view Score Distillation Sampling (SDS) and reconstruction losses based on the original observation. TeCH produces high-fidelity 3D clothed humans with consistent delicate texture, and detailed full-body geometry. Quantitative and qualitative experiments demonstrate that TeCH outperforms the state-of-the-art methods in terms of reconstruction accuracy and rendering quality. The code will be publicly available for research purposes at https://huangyangyi.github.io/TeCH

READ FULL TEXT

page 1

page 3

page 5

page 9

page 10

page 13

page 14

page 15

research
11/09/2022

ReFu: Refine and Fuse the Unobserved View for Detail-Preserving Single-Image 3D Human Reconstruction

Single-image 3D human reconstruction aims to reconstruct the 3D textured...
research
08/21/2023

TADA! Text to Animatable Digital Avatars

We introduce TADA, a simple-yet-effective approach that takes textual de...
research
08/21/2023

SCULPT: Shape-Conditioned Unpaired Learning of Pose-dependent Clothed and Textured Human Meshes

We present SCULPT, a novel 3D generative model for clothed and textured ...
research
05/25/2023

ZeroAvatar: Zero-shot 3D Avatar Generation from a Single Image

Recent advancements in text-to-image generation have enabled significant...
research
04/03/2023

DreamAvatar: Text-and-Shape Guided 3D Human Avatar Generation via Diffusion Models

We present DreamAvatar, a text-and-shape guided framework for generating...
research
07/15/2021

Single-image Full-body Human Relighting

We present a single-image data-driven method to automatically relight im...
research
04/06/2023

DITTO-NeRF: Diffusion-based Iterative Text To Omni-directional 3D Model

The increasing demand for high-quality 3D content creation has motivated...

Please sign up or login with your details

Forgot password? Click here to reset