DeepAI
Log In Sign Up

Learning elementary structures for 3D shape generation and matching

08/13/2019
by   Theo Deprelle, et al.
0

We propose to represent shapes as the deformation and combination of learnable elementary 3D structures, which are primitives resulting from training over a collection of shape. We demonstrate that the learned elementary 3D structures lead to clear improvements in 3D shape generation and matching. More precisely, we present two complementary approaches for learning elementary structures: (i) patch deformation learning and (ii) point translation learning. Both approaches can be extended to abstract structures of higher dimensions for improved results. We evaluate our method on two tasks: reconstructing ShapeNet objects and estimating dense correspondences between human scans (FAUST inter challenge). We show 16 shape reconstruction and outperform FAUST inter challenge state of the art by 6

READ FULL TEXT VIEW PDF

page 1

page 2

page 3

page 4

03/16/2022

Topology-Preserving Shape Reconstruction and Registration via Neural Diffeomorphic Flow

Deep Implicit Functions (DIFs) represent 3D geometry with continuous sig...
06/13/2018

Shape correspondences from learnt template-based parametrization

We present a new deep learning approach for matching deformable shapes b...
02/03/2022

Bending Graphs: Hierarchical Shape Matching using Gated Optimal Transport

Shape matching has been a long-studied problem for the computer graphics...
06/26/2020

Meta Deformation Network: Meta Functionals for Shape Correspondence

We present a new technique named "Meta Deformation Network" for 3D shape...
07/06/2019

Unsupervised cycle-consistent deformation for shape matching

We propose a self-supervised approach to deep surface deformation. Given...
09/06/2019

Real-time Deformation with Coupled Cages and Skeletons

Skeleton-based and cage-based deformation techniques represent the two m...
10/17/2018

A Convolutional Autoencoder Approach to Learn Volumetric Shape Representations for Brain Structures

We propose a novel machine learning strategy for studying neuroanatomica...

1 Introduction

Current surface-parametric approaches for generating a surface or aligning two surfaces, such as AtlasNet groueix2018 and 3D-CODED groueix2018b , rely on alignment of one or more shape primitives to a target shape. The shape primitives can be a set of patches or a sphere, as in AtlasNet, or a human template shape, as in 3D-CODED. These approaches could easily be extended to other parametric shapes, such as blocks roberts1963machine , generalized cylinders binford1971visual , or modern shape abstractions li2017grass ; sharma2018csgnet ; abstractionTulsiani17 . While surface-parametric approaches have achieved state-of-the-art results for (single-view) shape reconstruction groueix2018 and 3D shape correspondences groueix2018b , they rely on hand-chosen parametric shape primitives tuned for the target shape collection and task. In this paper, we ask – what is the right set of primitives to represent a collection of diverse shapes?

To address this question, we seek to go beyond manually choosing shape primitives and automatically learn what we call “learnable elementary structures” from a shape collection, which can be used for shape reconstruction and matching. The ability to automatically learn elementary structures allows the shape generator to find a better set of primitives for a shape collection and target task. We find that learned elementary structures correspond to recurrent parts among 3D objects. For example, in Figure 1, we show automatically learned elementary structures roughly corresponding to the tail, wing, and reactor of an airplane. Moreover, we find that learning the elementary structures leads to an improvement in shape reconstruction and correspondence accuracy.

We explore two approaches for learning elementary structures – patch deformation learning and point translation learning. For patch deformation learning, similar to AtlasNet groueix2018

, we start from a surface element, such as a 2D square, and deform it into the learned structure using a multi-layer perceptron  

rosenblatt1958perceptron . This approach has the advantage that the learned elementary structures are continuous surfaces. Its key difference with respect to AtlasNet is that the deformations, and thus the elementary structures, are common to all shapes. For point translation learning, starting from a fixed set of points, we optimize their position to reconstruct the target objects. The drawback of this approach is that it does not produce a continuous surface – only a finite set of points. However, this approach is more flexible since it can, for example, change the topology of the structure.

We show how to deform and combine our learnable elementary structures to explain a given 3D shape. At inference, given the learned elementary structures, we learn to position the structures by adjustment – a linear (projective) transformation will lead to maximum interpretability, while a complex transformation parameterized by a multi-layer perceptron will make our approaches generalizations of prior shape reconstruction methods groueix2018 ; groueix2018b using optimized instead of manually defined templates. Moreover, such representation allows for disentanglement of the structure’s shape and pose. We include structure learning in a deep architecture that unifies shape abstraction and deep surface deformation approaches.

We demonstrate that our architecture leads to improvements for 3D shape generation and matching – relative improvement over AtlasNet for generic object shape reconstruction and over 3D-CODED for human shape matching on Faust bogo2014cvpr , achieving state of the art for the latter task. Our code is available on our project webpage111http://imagine.enpc.fr/ deprellt/atlasnet2

(a) Input target shapes
(b) Learned elementary structures
(c) Our reconstructions
Figure 1: Problem statement. We seek to automatically learn a set of primitives (called “learned elementary structures”) for shape reconstruction and matching. (a) Input target shapes to reconstruct. (b) Learned elementary structures roughly corresponding to the tail, wing, and reactor of airplanes. (c) Our output reconstructions with learned elementary structures highlighted.

2 Related Work

Primitive fitting is a classic topic in computer vision 

roberts1963machine , with a large number of methods targeting parsimonious shape approximations, such as generalized cylindersbinford1971visual and geons biederman1987recognition . Efficient fitting of these primitives attracted a lot of research efforts kaiser2018survey ; li_globFit_sigg11 ; schnabel-2009-completion ; schnabel2007efficient . Since these methods analyze shapes independently, they are not expected to use the primitives consistently across different objects, which makes the result unsuitable for discovering a common structure in a collection of shapes, performing consistent segmentation, or correspondence estimation. To address these limitations some methods optimize for consistent primitive fitting over the entire shape collection Kim13 , or aim to discover a consistent set of parts Golovinskiy09 ; Huang11 ; Sidi2011

. The resulting optimization problems are usually non-convex, and thus existing solutions tend to be slow, require heuristics, and are prone to being stuck in local optima.

Learning-based techniques offer a promising alternative to hand-crafted heuristics. Zhu et al. zou20173d

use a Recurent Neural Network supervised by a traditional heuristic-based algorithm for cuboid fitting. Tulsiani

et al. abstractionTulsiani17 use reconstruction loss to predict parameters of the cuboids that approximate an input shape, and thus do not require any direct supervision. Several recent techniques, concurrent to our work, extend this approach by using more complex primitives that can better approximate the surface, such as anisotropic 3D Gaussians genova2019learning , categorie specifique morphable model  cmrKanazawa18 or superquadrics Paschalidou2019CVPR . All of these techniques use a collection of simple hand-picked parametric primitives. In contrast, we propose to learn a set of deformable primitives that best approximate a collection of shapes.

One can further improve reconstruction by fitting a diverse set of primitives li2018supervised or constructive solid geometry graphs sharma2018csgnet . These methods, however, usually do not produce consistent fitting across different shapes, and thus cannot be used to discover common shape structures or inter-shape relationships.

On the other side of the spectrum, instead of simple primitives, some techniques fit deformable mesh models Allen02 ; Allen03 ; Loper15 ; Zuffi15

. While they can capture complex structures, these techniques are also prone to being stuck in local optima, due to large number of degrees of freedom (e.g., mesh vertex coordinates).

Neural network architectures have been used to facilitate the mesh fitting groueix2018b , learning to predict the deformation of a template to reconstruct unstructured input point cloud. This approach is sensitive to the choice of the template. We demonstrate that our method improves the quality of the fitting by learning the structure of the reference shape. Neural mesh fitting has been also employed for geometrically and topologically diverse datasets that do not have a natural template. In these cases, meshed planes or spheres can be deformed into complex 3D structures groueix2018 ; Yang_2018 . We extend this line of work by proposing a technique for learning the base shapes that are further used to approximate the shapes in the collection. Learning these elementary structures enables us to more accurately and consistently reconstruct the shapes in the collection.

3 Approach

(a) Points translation learning module
(b) Patch deformation learning module
(c) Elementary structures combination model
Figure 2: Approach overview. At training time, we learn (a) translations or (b) deformations that transform points from the unit square into shared learned elementary structures. (c) At evaluation time, we transform each elementary structure to target shape using learned shape-dependent adjustment networks that produce points on the surface of the output shape .

We aim to learn shared elementary structures to reconstruct a set of 3D shapes. We visualize an overview of our approach in Figure 2. We formulate two ways to learn elementary structures – via patch deformation learning and point translation learning modules. The elementary structures are learned over the entire training set and do not depend on the input during testing. At test time, the elementary structures are deformed by adjustment modules to create the output 3D shape. These modules take as inputs features computed from the input via an encoder network and the coordinates of the elementary structure points and output the 3D coordinates of the deformed primitives.

For the task of 3D shape reconstruction, we assume that we are given a training set of target shapes . Our goal is to reconstruct the target shapes using a set of learned elementary structures , which are deformed via shape-dependent adjustment modules

. We represent each shape by a feature vector

computed by a point set encoder (defined later in this section). Each adjustment module takes as inputs the coordinates of a point in the associated elementary structure and the feature vector of the target shape and outputs 3D coordinates of the corresponding point. The output shape can thus be written as the union over learned and adjusted elementary structures,

(1)

If the elementary structures were unit squares or a unit sphere, then this equation would describe exactly the AtlasNet groueix2018 model. On the other hand, the 3D-CODED model groueix2018b uses an instance of as a single elementary structure. Generalizing these approaches, our goal is to automatically learn the elementary structures over a shape collection. The intuition behind our approach is that if the elementary structures have useful shapes to reconstruct the target, the adjustment should be easier to learn and more interpretable.

3.1 Learnable elementary structures

For each , we start from an initial surface on which we sample points to obtain an initial point cloud . We then pass each sampled point for through elementary structure learning modules .

We consider two types of elementary structure learning module .

The first type, patch deformation learning module, learns a continuous mapping to obtain deformed points starting from sampled point . The intuition behind the deformation module is that elementary structures should be surface elements, and can thus be deduced from the transformation of the original surfaces . Alternatively, we consider a point translation learning module which translates independently each of the points by a learned vector , . This module thus allows the network to update independently the position of each point on the surface. The result of either module results in a set of elementary structure points , and we write the elementary structure as the union of the independently deformed or translated points .

In Section 4 we will show that different choices here can be desirable depending on the application domain.

Dimensionality of the elementary structures.

While it is natural to consider elementary structures as sets of 3D points, we can extend the idea to other dimensions. We experimented with 2D, 3D, and 10D elementary structures and show that while they are less interpretable, higher-dimensional structures lead to better shape reconstruction results.

3.2 Architecture details

The following describes more details of our final network.

Shape encoder.

We represent the input shape as a point cloud, and we use as shape encoder a simplified version of the PointNet network qi2017pointnet used in groueix2018b ; groueix2018

. We represent each 3D point of the input shape as a 1024 dimensional vector using a multi-layer perceptron with 3 hidden layers of 64, 128 and 1024 neurons and ReLU activations. We then apply max-pooling over all point features followed by a linear layer, producing a global shape feature used as input to the adjustment modules.

Patch deformation learning module.

The patch deformation learning modules are continuous-space deformations that we learn as multi-layer perceptrons with 3 hidden layers of 128, 128 and 3 neurons and ReLU activations. This module takes as input coordinates of points in the initial structures and can compute not only a set of points  groueix2018 but the full image of a surface. If this module is used, we can densely sample points on the generated surface.

Point translation learning module.

The point translation learning modules learn a translation for each of the points of the associated initial structure. While this step gives more flexibility than generating points through the patch deformation learning module, it can only be applied for a fixed number of points, similar to point-based shape generation fan2017point .

Adjustment module.

The goal of the adjustment modules is to reconstruct the input shape by positioning each elementary structure. The intuition is that this adjustment should be relatively simple. However, we can expect the quality of the reconstruction to increase using more complex adjustment modules. In this paper, we consider two cases:

  • Linear adjustment: each adjustment module applies an affine transformation to the corresponding elementary structure. The parameters of this transformation are predicted by a multi-layer perceptron that takes as input the point cloud feature vector generated by the encoder. We use three hidden MLP layers (512, 512, 12), ReLU activation, BatchNorm layers and a hyperbolic tangent at the last layer for this module.

  • MLP adjustment: each adjustment module uses a multi-layer perceptron (MLP) that takes as inputs the concatenation of the coordinates of a point from the associated elementary structure and the shape feature predicted by the shape encoder and outputs 3D coordinates. We use the same architecture as groueix2018 for this network to obtain comparable results.

3.3 Losses and training

We now discuss two scenarios in which we tested our approach.

Training with correspondences.

In this scenario, we assume point correspondences across all training examples and a common template that we can use as an initial structure for all shapes. More precisely, we assume that each training shape is represented as an ordered set of 3D points in consistent locations on all shapes. Since all shapes are in correspondence, we consider a single elementary structure () and N sampled points on the shape . We then train our network to minimize the following squared loss between sampled points on each training shape to reconstructed points starting from sampled template points :

(2)

where are the parameters of the networks. Note that at inference, we do not need to know the correspondences of the points in the test shape, since they are processed by the point set encoder which is invariant to the order of the points. Instead, the points in the reconstruction shapes will be in correspondence with the elementary structure and by extension with each other. We use this property to predict correspondences between test shapes, following the pipeline of groueix2018b . Learning the elementary structures is the difference between our approach and 3D-CODED groueix2018b in this scenario, which leads to improved reconstruction and correspondence accuracy.

Training without correspondences.

We are also able to train our system when no correspondence supervision is available during training. In this case, there are many options for our choice of elementary structures. To be comparable with AtlasNet groueix2018 , we will assume we have elementary structures and that each initial structure is a unit 2D square patch. For a given training shape , we compute the output shape according to Equation 1, and train our network’s parameters to minimize the symmetric Chamfer distance fan2017point between the point clouds and .

(3)

where are the parameters of the networks. In all of our experiments, we used .

(a) Category-specific 2D elementary structures (3 out of 10 structures) learned for chairs (left) and plane (righ).
(b) Reconstructions using elementary structures with category-specific training.
(c) 2D elementary structure learned from all categories (7 out of 10 structures are shown).
(e) AtlasNet reconstruction using square patch primitives trained on all categories
(d) Our reconstructions using 2D elementary structures trained on all categories.
(d) Our reconstructions using 2D elementary structures trained on all categories.
Figure 3: We visualize elementary structures using point learning and MLP adjustment modules. For all reconstruction results, we show in color the points corresponding to the visualized 2D primitives. For AtlasNet, the primitives are unit squares (so we do not show the elementary structures), and we visualize seven of them for the reconstruction (similarly to our method). Contrary to AtlasNet, our learned elementary structures have limited overlap in the reconstructions and better reconstructs the shapes.

max width=0.6 Single-category training Multi-category training Airplanes Chairs Airplanes Chairs All Linear adjustment AtlasNet groueix2018 1.57 4.14 2.22 3.72 3.07 Deformation 1.16 2.76 1.49 2.52 2.26 Points 1.04 2.00 1.35 2.47 2.11 MLP adjustment AtlasNet groueix2018 0.91 1.64 0.81 1.50 1.45 Deformation 0.87 1.56 0.81 1.25 1.43 Points 0.79 1.43 0.71 1.25 1.22  max width=0.30 Multi-category training Points Def. MLP adjustment 2D 1.28 1.42 3D 1.22 1.43 10D 1.21 1.39 Linear adjustment 2D 2.45 2.75 3D 2.11 2.26 10D 1.66 1.90

Table 1: ShapeNet reconstruction. We evaluate variants of our method for single- and multi-category reconstruction tasks. Left: Linear vs MLP adjustment, Patch Deformation vs Points Translation with 3D elementary structures. Right: different template dimensionality and deformation vs points learning modules in the multi-category setup with MLP-adjustement. We report Chamfer distance (multiplied by ). AtlasNet uses 10 patch primitives, which is the same as our approach, without the learned elementary structures.

Training details.

We use the Adam optimizer with a learning rate of 0.001, a batch size of 16, and batch normalization layers. We train our method using input point clouds of 2500 points when correspondences are not available and 6800 points when correspondences are available. When training using only the deformation modules

, we resample the initial surfaces at each training step to minimize over-fitting. At inference time, we sample a regular grid to allow easy mesh generation. We train our model on an NVIDIA 1080Ti GPU, with a 16 core Intel I7-7820X CPU (GHz), 126GB RAM and SSD storage. Training takes about 48h for most experiments. Using the trained models from the official implementation on all categories, AtlasNet-25 performance is (see also Table 1 in the Atlasnet paper). Using the released code to train AtlasNet-10 yields an error of . By adding a learning rate schedule to the original implementation we decreased this error to and report this improved baseline (see Table 1).

4 Experiments

In this section, we show qualitative and quantitative results of our approach on the tasks of shape reconstruction and shape matching.

4.1 Generic object shape reconstruction

We evaluate our approach on non-articulated generic 3D object shapes for the task of shape reconstruction. We use the training setting without correspondences described in Section 3.3.

Dataset, evaluation criteria, baseline.

We evaluate on the ShapeNet Core dataset ChangFGHHLSSSSX15 . For single-category reconstruction, we evaluated over airplane (5424/1360 train/test shapes) and chair (3248/816) categories. For multi-category reconstruction, we used 13 categories – airplane, bench, cabinet, car, chair, monitor, lamp, speaker, firearm, couch, table, cellphone, watercraft (31760/7952). We report the symmetric Chamfer distance between the reconstructed and target shapes. All reported Chamfer results are multiplied by . As a baseline, we compare against AtlasNet groueix2018 with ten unit-square primitives.

Single-category shape reconstruction.

For our first experiment, we trained separate networks for the different ShapeNet Core categories. Figure 2(a) demonstrates learned 2D elementary structures using ten 2D unit squares as initial structures . In Figure  2(b), we show shape reconstructions using our points translation learning module with MLP adjustments. Note the emergence of symmetric and topologically complex elementary structures.

(a) MLP adjustment
(b) Linear adjustment
Figure 4: Three (out of ten) learned 3D elementary structures learned by the point translation learning approach when training on all ShapeNet categories.
Figure 5: 3D elementary structure obtained with point learning when initializing the training from a template shape (left) or a random set of points (right). See text for details.

Multi-class shape reconstruction.

We now evaluate how well our method generalizes when trained on multiple categories, again using 2D elementary structures with point translation learning module and MLP-adjustements. As in single-category case, we observe discovery of non-trivial 2D elementary structures (Figure 2(c)) that are used to accurately reconstruct the shapes (Figure 3), with higher fidelity than the baseline performance of AtlasNet with ten 2D square patches (Figure 2(e)). Note how AtlasNet is less faithful to the topology of reconstructed shapes, incorrectly synthesizing geometry in hollow areas between the back and the seat. Our quantitative evaluation in Table 1 confirms that AtlasNet provides less accurate reconstructions than our method.

Linear vs MLP adjustment.

We evaluated networks trained in both the single- and multi-category settings with linear and MLP adjustment modules using 3D learned elementary structures (Table 1 left, Figure 4

). In all experimental setups, we observe that the MLP adjustment offers significant quantitative improvements over restricting the network to use linear transformations of the elementary structures. This result is expected as linear adjustment allows only limited adaptation of the elementary structures for each shape. Similar to shape abstraction methods

abstractionTulsiani17 , linear adjustment allows a better intuition of the shape generation process but limits the reconstruction accuracy. Using MLP adjustments, however, offers the network more flexibility to faithfully reconstruct the shapes.

Patch deformation vs points translation modules.

We compare using patch deformation vs points translation modules in Table 1. The patch deformation learning module does not allow topological changes and discontinuities in mapping, and produces inferior results in comparison to points translation learning. On the other hand, learning patch deformations enables the estimation of the entire deformation field. Thus one can warp an arbitrary number of points or even tessellate the domain and warp the entire mesh to generate the polygonal surface, which is more amenable to tasks such as rendering.

Higher-dimensional structures.

We experimented with the dimensionality of the learned elementary structures. Figures 2(a) and 2(c) suggest that learned 2D elementary structures can capture interesting topological and symmetric aspects of the data – splitting, for instance, the patch into two identical parts for the legs of the chairs. note also the variable point density. Similarly, learned 3D elementary structures with linear adjustment and patch deformation learning modules are shown in Figure 1 for the airplane category. Note that they roughly correspond to meaningful parts, such as wings, tail and reactor. Figure 4 shows 3D elementary structures inferred from all ShapeNet categories, where the learned structures include non-trivial elements such as symmetric planes, sharp angles, and smooth parabolic surfaces. The learned structures are often correspond to consistent parts in the reconstructions. In our quantitative evaluations (Table 1, right) we found that the results improve with the dimensionality. The improvement diminishes for higher-dimensional spaces and are more difficult to visualize and interpret.

Consistency in template elementary structures.

We experimented with several initializations of our elementary structures on the ShapeNet plane category. We used the point translation learning method and a single 3D elementary structure. In Figure 5, we show our results when initializing the elementary structure with either a plane 3D model (left) or a set of random 3D points sampled uniformly (right).

Chairs Table
AtlasNet
Patch.
Point.
Figure 6: Category generalization. Chamfer distance for networks trained on chairs and tested on either the chairs or tables test sets.

Notice that the learned 3D elementary structure is similar regardless of the initial template shape.

Generalization to new categories.

To test the generality of our approach, we trained on the chair category using ten 2D elementary structures and tested on the table category. As shown in Figure 6, point translation learning outperforms both patch deformation learning and AtlasNet. Figure 7 shows qualitatively how the elementary structures are positioned on chairs and tables. Notice how the chair and table legs are reconstructed by the same elementary structures.

Figure 7: Elementary structures learned on chairs (left) used to reconstruct chairs and tables (right).
Param. Chamfer
AtlasNet
6-layer AN
Patch.
Point.
Figure 8: Impact of number of parameters on reconstruction error.

Number of parameters.

In Figure 8, we show the number of parameters for AtlasNet and our method. Our method has less than 1% additional parameters to learn the elementary structures – and for patch deformation and point translation, respectively (orders of magnitude smaller than for the full network). During inference, our approach has the same complexity as AtlasNet as the elementary structures are precomputed and remain fixed for all shapes. We also tried training AtlasNet with six layers (6-layer AN), which significantly increases the number of parameters. Our approach with points translation learning outperforms all methods.

4.2 Human shape reconstruction and matching

We now evaluate our approach on 3D human shapes for the tasks of shape reconstruction and matching using the training setup with correspondences described in Section 3.3. For this task, we use a single elementary structure for the human body using one of the meshes as the initial structure . Since we use a single elementary structure and the shapes are deformable, we only report results using the MLP-adjustment.

Datasets, evaluation criteria, baselines.

We train our method using the SURREAL dataset  varol17_surreal , extended to include some additional bend-over poses as in 3D-CODED groueix2018b . We use 229,984 SURREAL meshes of humans in various poses for training and 224 SURREAL meshes to test reconstruction quality. To evaluate correspondences on real data, we use the FAUST benchmark bogo2014cvpr consisting of 200 testing scans with vertices from the “inter” challenge, including noise and holes which are not present in our training data. As a baseline, we compared against 3D-CODED groueix2018b .

(a) Learned points
(b) Learned deformation
(c) Learned deformation
Figure 9: Initial shape (left) and learned elementary structure (right) using the deformation or points learning modules. Notice the similarity between the elementary structure learned with the different approaches.

Results.

Figure 9 shows learned elementary structures using deformation or points translation learning and different initial surfaces. We observe that the learned templates are inflated, bent, and with their arm and legs in a similar pose, suggesting a reasonable amount of consistency in the properties of a desirable primitive shape for this task.

As before, we found that points translation learning provides the best reconstruction (see SURREAL column in Table 2). Both of our approaches also provide lower reconstruction loss than 3D-CODED.

We used reconstruction to estimate correspondences by finding closest points on the deformed elementary structure as in 3D-CODED groueix2018b . We report correspondence error in the “FAUST” column in Table 2. We observe that deformation learning provides better correspondences than points learning, also yielding state-of-the-art results and clear improvement over 3D-CODED. This result is not surprising because understanding the deformation field for the entire surface is more relevant for matching and correspondence problems.

SURREAL varol17_surreal FAUST bogo2014cvpr
3D-CODED 1.33 2.96
Deformation 1.17 2.76
Points 1.11 3.05
SURREAL varol17_surreal FAUST bogo2014cvpr
Points Deform. Points Deform.
2D 1.55 1.72 3.28 3.31
3D 1.11 1.17 3.05 2.76
10D 1.01 1.01 2.85 2.77
Table 2: Human correspondences and reconstruction. We evaluate different variants of our method (with deformation vs points translation learning and different template dimensionality) for surface reconstruction (SURREAL column) and matching (FAUST column). We report Chamfer loss for the former and correspondence error for the latter (measured by the distance between corresponding points). Results in the left table are with 3D elementary structures, and the only difference with the 3D-CODED baseline is thus the template/elementary structure learning. The table on the right shows results with elementary structures of different dimensions.

Elementary structure dimension.

Similar to generic object reconstruction, we evaluate with 2D, 3D and 10D elementary structures (Table 2, right). Note that when using the patch deformation learning module we control the output size and therefore it is easy to map the input 3D template to higher- or lower-dimensional elementary structure. On the other hand the points translation learning module does not allow to change dimensionality of the input template. Hence, for 2D elementary structures we project the 3D template (front-facing human in a T-pose) to a front plane, and for 10D elementary structures we embed the 3D human into a hyper-cube, keeping higher dimensions as zero. The difference in performance is clearer for human reconstruction than for generic object reconstruction, which can be related both to the fact that humans are complex with articulations and that we use a single elementary structure for all human reconstructions.

5 Conclusion

We have presented a method to take a collection of training shapes and learned common elementary structures that can be deformed and composed to consistently reconstruct arbitrary shapes. We learn consistent structures without explicit point supervision between shapes and we demonstrate that using our structures for reconstruction and correspondence tasks results in significant quantitative improvements. When trained on shape categories, these structures are often interpretable. Moreover, our deformation learning approach learns elementary structures as the deformation of continuous surfaces, resulting in output surfaces that can densely sampled and meshed at test time. Our approach opens up possibilities for other applications, such as shape morphing and scan completion.

Acknowledgments.

This work was partly supported by ANR project EnHerit ANR-17-CE23-0008, Labex Bézout, and gifts from Adobe to École des Ponts.

References

  • (1) B. Allen, B. Curless, and Z. Popovic. Articulated body deformation from range scan data. SIGGRAPH, 2002.
  • (2) B. Allen, B. Curless, and Z. Popovic. The space of human body shapes: reconstruction and parameterization from range scans. SIGGRAPH, 2003.
  • (3) I. Biederman. Recognition-by-components: a theory of human image understanding. Psychological review, 94(2):115, 1987.
  • (4) I. Binford. Visual perception by computer. In IEEE Conference of Systems and Control, 1971.
  • (5) F. Bogo, J. Romero, M. Loper, and M. J. Black. Faust: Dataset and evaluation for 3d mesh registration. In CVPR, 2014.
  • (6) A. X. Chang, T. A. Funkhouser, L. J. Guibas, P. Hanrahan, Q. Huang, Z. Li, S. Savarese, M. Savva, S. Song, H. Su, J. Xiao, L. Yi, and F. Yu. Shapenet: An information-rich 3d model repository. CoRR, abs/1512.03012, 2015.
  • (7) H. Fan, H. Su, and L. J. Guibas. A point set generation network for 3d object reconstruction from a single image. In

    Proceedings of the IEEE conference on computer vision and pattern recognition

    , pages 605–613, 2017.
  • (8) K. Genova, F. Cole, D. Vlasic, A. Sarna, W. T. Freeman, and T. A. Funkhouser. Learning shape templates with structured implicit functions. CoRR, abs/1904.06447, 2019.
  • (9) A. Golovinskiy and T. Funkhouser. Learning Consistent Segmentation of 3D Models. Computers and Graphics (Shape Modeling International), 2009.
  • (10) T. Groueix, M. Fisher, V. G. Kim, B. Russell, and M. Aubry. 3d-coded : 3d correspondences by deep deformation. In ECCV, 2018.
  • (11) T. Groueix, M. Fisher, V. G. Kim, B. Russell, and M. Aubry. AtlasNet: A Papier-Mâché Approach to Learning 3D Surface Generation. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2018.
  • (12) Q. Huang, V. Koltun, and L. Guibas.

    Joint-Shape Segmentation with Linear Programming.

    ACM Transactions on Graphics (Proc. SIGGRAPH Asia), 2011.
  • (13) A. Kaiser, J. A. Ybanez Zepeda, and T. Boubekeur. A survey of simple geometric primitives detection methods for captured 3d data. In Computer Graphics Forum. Wiley Online Library, 2018.
  • (14) A. Kanazawa, S. Tulsiani, A. A. Efros, and J. Malik. Learning category-specific mesh reconstruction from image collections. In ECCV, 2018.
  • (15) V. G. Kim, W. Li, N. J. Mitra, S. Chaudhuri, S. DiVerdi, and T. Funkhouser. Learning Part-based Templates from Large Collections of 3D Shapes. Transactions on Graphics (Proc. of SIGGRAPH), 32(4), 2013.
  • (16) J. Li, K. Xu, S. Chaudhuri, E. Yumer, H. Zhang, and L. Guibas.

    Grass: Generative recursive autoencoders for shape structures.

    ACM Transactions on Graphics (TOG), 36(4):52, 2017.
  • (17) L. Li, M. Sung, A. Dubrovina, L. Yi, and L. Guibas. Supervised fitting of geometric primitives to 3d point clouds. arXiv preprint arXiv:1811.08988, 2018.
  • (18) Y. Li, X. Wu, Y. Chrysanthou, A. Sharf, D. Cohen-Or, and N. J. Mitra. Globfit: Consistently fitting primitives by discovering global relations. ACM Transactions on Graphics, 30(4):52:1–52:12, 2011.
  • (19) M. Loper, N. Mahmood, J. Romero, Pons-Moll, and M. J. G., Black. Smpl: A skinned multi-person linear model. SIGGRAPH Asia, 2015.
  • (20) D. Paschalidou, A. O. Ulusoy, and A. Geiger. Superquadrics revisited: Learning 3d shape parsing beyond cuboids. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), June 2019.
  • (21) C. R. Qi, H. Su, K. Mo, and L. J. Guibas.

    Pointnet: Deep learning on point sets for 3d classification and segmentation.

    In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 652–660, 2017.
  • (22) L. G. Roberts. Machine perception of three-dimensional solids. PhD thesis, Massachusetts Institute of Technology, 1963.
  • (23) F. Rosenblatt. The perceptron: a probabilistic model for information storage and organization in the brain. Psychological review, 65(6):386, 1958.
  • (24) R. Schnabel, P. Degener, and R. Klein. Completion and reconstruction with primitive shapes. Computer Graphics Forum (Proc. of Eurographics), 28(2):503–512, Mar. 2009.
  • (25) R. Schnabel, R. Wahl, and R. Klein. Efficient ransac for point-cloud shape detection. In Computer graphics forum, volume 26, pages 214–226. Wiley Online Library, 2007.
  • (26) G. Sharma, R. Goyal, D. Liu, E. Kalogerakis, and S. Maji. Csgnet: Neural shape parser for constructive solid geometry. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5515–5523, 2018.
  • (27) O. Sidi, O. van Kaick, Y. Kleiman, H. Zhang, and D. Cohen-Or.

    Unsupervised co-segmentation of a set of shapes via descriptor-space spectral clustering.

    In Transactions on Graphics (Proc. of SIGGRAPH Asia), pages 126:1–126:10, 2011.
  • (28) S. Tulsiani, H. Su, L. J. Guibas, A. A. Efros, and J. Malik. Learning shape abstractions by assembling volumetric primitives. In Computer Vision and Pattern Recognition (CVPR), 2017.
  • (29) G. Varol, J. Romero, X. Martin, N. Mahmood, M. J. Black, I. Laptev, and C. Schmid. Learning from synthetic humans. In CVPR, 2017.
  • (30) Y. Yang, C. Feng, Y. Shen, and D. Tian. Foldingnet: Point cloud auto-encoder via deep grid deformation. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun 2018.
  • (31) C. Zou, E. Yumer, J. Yang, D. Ceylan, and D. Hoiem.

    3d-prnn: Generating shape primitives with recurrent neural networks.

    In Proceedings of the IEEE International Conference on Computer Vision, pages 900–909, 2017.
  • (32) S. Zuffi and M. J. Black. The stitched puppet: A graphical model of 3d human shape and pose. Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2015.