## 1 Introduction

We live in a three-dimensional world, and a proper understanding of its volumetric structure is crucial for acting and planning. However, we perceive the world mainly via its two-dimensional projections. Based on these projections, we are able to infer the three-dimensional shapes and poses of the surrounding objects. How does this volumetric shape perception emerge from observing only from two-dimensional projections? Is it possible to design learning systems with similar capabilities?

Deep learning methods have recently shown promise in addressing these questions (Yan2016; Tulsiani2017drc). Given a set of views of an object and the corresponding camera poses, these methods learn 3D shape via the reprojection error: given an estimated shape, one can project it to the known camera views and compare to the provided images. The discrepancy between these generated projections and the training samples provides training signal for improving the shape estimate. Existing methods of this type have two general restrictions. First, these approaches assume that the camera poses are known precisely for all provided images. This is a practically and biologically unrealistic assumption: a typical intelligent agent only has access to its observations, not its precise location relative to objects in the world. Second, the shape is predicted as a low-resolution (usually voxels) voxelated volume. This representation can only describe very rough shape of an object. It should be possible to learn finer shape details from 2D supervision.

In this paper, we learn high-fidelity shape models solely from their projections, without ground truth camera poses. This setup is challenging for two reasons. First, estimating both shape and pose is a chicken-and-egg problem: without a good shape estimate it is impossible to learn accurate pose because the projections would be uninformative, and vice versa, an accurate pose estimate is necessary to learn the shape. Second, pose estimation is prone to local minima caused by ambiguity: an object may look similar from two viewpoints, and if the network converges to predicting only one of these in all cases, it will not be able to learn predicting the other one. We find that the first problem can be solved surprisingly well by joint optimization of shape and pose predictors: in practice, good shape estimates can be learned even with relatively noisy pose predictions. The second problem, however, leads to drastic errors in pose estimation. To address this, we train a diverse ensemble of pose predictors and distill those to a single student model.

To allow learning of high-fidelity shapes, we use the point cloud representation, in contrast with voxels used in previous works. Point clouds allow for computationally efficient processing, can produce high-quality shape models (Fan2017), and are conceptually attractive because they can be seen as “matter-centric”, as opposed to “space-centric” voxel grids. To enable learning point clouds without explicit 3D supervision, we implement a differentiable projection operator that, given a point set and a camera pose, generates a 2D projection – a silhouette, a color image, or a depth map. We dub the formulation “Differentiable Point Clouds”.

We evaluate the proposed approach on the task of estimating the shape and the camera pose from a single image of an object. The method successfully learns to predict both the shape and the pose, with only a minor performance drop relative to a model trained with ground truth camera poses. The point-cloud-based formulation allows for effective learning of high-fidelity shape models when provided with images of sufficiently high resolution as supervision. We demonstrate learning point clouds from silhouettes and augmenting those with color if color images are available during training. Finally, we show how the point cloud representation allows to automatically discover semantic correspondences between objects.

## 2 Related Work

Reconstruction of three-dimensional shapes from their two-dimensional projections has a long history in computer vision, constituting the field of 3D reconstruction. A review of this field goes outside of the scope of this paper; however, we briefly list several related methods.

Cashman2013 use silhouettes and keypoint annotation to reconstruct deformable shape models from small class-specific image collections, Vicente2014 apply similar methods to a large-scale Pascal VOC dataset, Tulsiani2017deformable reduce required supervision by leveraging computer vision techniques. These methods show impressive results even in the small data regime; however, they have difficulties with representing diverse and complex shapes. Loper2014 implement a differentiable renderer and apply it for analysis-by-synthesis. Our work is similar in spirit, but operates on point clouds and integrates the idea of differentiable rendering with deep learning. The approach of Rhodin2015 is similar to our technically in that it models human body with a set of Gaussian density functions and renders them using a physics-motivated equation for light transport. Unlike in our approach, the representation is not integrated into the learning framework and requires careful initial placement of the Gaussians, making it unsuitable for automated reconstruction of arbitrary shape categories. Moreover, the projection method scales quadratically with the number of Gaussians, which limits the maximum fidelity of the shapes being represented.Recently the task of learning 3D structure from 2D supervision is being addressed with deep-learning-based methods. The methods are typically based on reprojection error – comparing 2D projections of a predicted 3D shape to the ground truth 2D projections. Yan2016 learn 3D shape from silhouettes, via a projection operation based on selecting the maximum occupancy value along a ray. Tulsiani2017drc

devise a differentiable formulation based on ray collision probabilities and apply it to learning from silhouettes, depth maps, color images, and semantic segmentation maps.

Lin2018represent point clouds by depth maps and re-project them using a high resolution grid and inverse depth max-pooling. Concurrently with us,

Kato2018 propose a differentiable renderer for meshes and use it for learning mesh-based representations of object shapes. All these methods require exact ground truth camera pose corresponding to the 2D projections used for training. In contrast, we aim to relax this unrealistic assumption and learn only from the projections.Rezende2016 explore several approaches to generative modeling of 3D shapes based on their 2D views. One of the approaches does not require the knowledge of ground truth camera pose; however, it is only demonstrated on a simple dataset of textured geometric primitives. Most related to our submission is the concurrent work of Tulsiani2018. The work extends the Differentiable Ray Consistency formulation Tulsiani2017drc

to learning without pose supervision. The method is voxel-based and deals with the complications of unsupervised pose learning using reinforcement learning and a GAN-based prior. In contrast, we make use of a point cloud representation, use an ensemble to predict the pose, and do not require a prior on the camera poses.

The issue of representation is central to deep learning with volumetric data. The most commonly used structure is a voxel grid - a direct 3D counterpart of a 2D pixelated image (Choy2016; Wu2016). This similarity allows for simple transfer of convolutional network architectures from 2D to 3D. However, on the downside, the voxel grid representation leads to memory- and computation-hungry architectures. This motivates the search for alternative options. Existing solutions include octrees (Tatarchenko2017), meshes (Kato2018; Yang2018), part-based representations (Tulsiani2017primitives; Li2017), multi-view depth maps (Soltani2017), object skeletons (Wu2018), and point clouds (Fan2017; Lin2018). We choose to use point clouds in this work, since they are less overcomplete than voxel grids and allow for effective networks architectures, but at the same time are more flexible than mesh-based or skeleton-based representations.

## 3 Single-view Shape and Pose Estimation

We address the task of predicting the three-dimensional shape of an object and the camera pose from a single view of the object. Assume we are given a dataset of views of objects, with views available for the -th object: . Here denotes a color image and – the projection of some modality (silhouette, depth map of a color image) from the same view. Each view may be accompanied with the corresponding camera pose , but the more interesting case is when the camera poses are not known. We focus on this more difficult scenario in the remainder of this section.

An overview of the model is shown in Figure 1. Assume we are given two images and of the same object. We use parametric function approximators to predict a 3D shape (represented by a point cloud) from one of them , and the camera pose from the other one: . In our case, and

are convolutional networks that share most of their parameters. Both the shape and the pose are predicted as fixed-length vectors using fully connected layers.

Given the predictions, we render the predicted shape from the predicted view: , where denotes the differentiable point cloud renderer described in Section 4

. The loss function is then the discrepancy between this predicted projection and the ground truth. We use standard MSE in this work both for all modalities, summed over the whole dataset:

(1) |

Intuitively, this training procedure requires that for all pairs of views of the same object, the renderings of the predicted point cloud match the provided ground truth views.

(a) Pose ambiguity | (b) Training an ensemble of pose regressors |

Estimating pose with a distilled ensemble. We found that the basic implementation described above fails to predict accurate poses. This is caused by local minima: the pose predictor converges to either estimating all objects as viewed from the back, or all viewed from the front. Indeed, based on silhouettes, it is difficult to distinguish between certain views even for a human, see Figure 2 (a).

To alleviate this issue, instead of a single pose regressor , we introduce an ensemble of pose regressors (see Figure 2 (b)) and train the system with the “hindsight” loss (Guzman-rivera2012; Chen2017):

(2) |

The idea is that each of the predictors learns to specialize on a subset of poses and together they cover the whole range of possible values. No special measures are needed to ensure this specialization: it emerges naturally as a result of random weight initialization if the network architecture is appropriate. Namely, the different pose predictors need to have several (at least , in our experience) non-shared layers.

In parallel with training the ensemble, we distill it to a single regressor by using the best model from the ensemble as the teacher. This best model is selected based on the loss, as in Eq. (2). At test time we discard the ensemble and use the distilled regressor to estimate the camera pose. The loss for training the student is computed as an angular difference between two rotations represented by quaternions: , where denotes the real part of the quaternion. We found that standard MSE loss performs poorly when regressing rotation.

Network architecture. We implement the shape and pose predictor with a convolutional network with two branches. The network starts with a convolutional encoder with a total of layers,

of which have stride

. These are followed by shared fully connected layers, after which the network splits into two branches for shape and pose prediction. The shape branch is an MLP with one hidden layer. The point cloud of points is predicted as a vector with dimensionality (point positions) or (positions and RGB values). The pose branch is an MLP with one shared hidden layer and two more hidden layers for each of the pose predictors. The camera pose is predicted as a quaternion. In the ensemble model we use pose predictors. The “student” model is another branch with the same architecture.## 4 Differentiable Point Clouds

A key component of our model is the differentiable point cloud renderer . Given a point cloud and a camera pose , it generates a view . The point cloud may have a signal, such as color, associated with it, in which case the signal can be projected to the view.

The high-level idea of the method is to smooth the point cloud by representing the points with density functions. Formally, we assume the point cloud is a set of tuples , each including the point position , the size parameter , and the associated signal (for instance, an RGB color). In most of our experiments the size parameter is a two-dimensional vector including the covariance of an isotropic Gaussian and a scaling factor. However, in general can represent an arbitrary parametric distribution: for instance, in the supplement we show experiments with Gaussians with a full covariance matrix. The size parameters can be either specified manually or learned jointly with the point positions.

The overall differentiable rendering pipeline is illustrated in Figure 3. For illustration purposes we show 2D-to-1D projection in the figure, but in practice we perform 3D-to-2D projection. We start by transforming the positions of points to the standard coordinate frame by the projective transformation corresponding to the camera pose of interest: . The transform accounts for both extrinsic and intrinsic camera parameters. We also compute the transformed size parameters (the exact transformation rule depends on the distribution used). We set up the camera transformation matrix such that after the transform, the projection amounts to orthogonal projection along the third axis.

To allow for the gradient flow, we represent each point by a smooth function . In this work we set to scaled Gaussian densities. The occupancy function of the point cloud is a clipped sum of the individual per-point functions:

(3) |

where are the size parameters. We discretize the resulting function to a grid of resolution . Note that the third index corresponds to the projection axis, with index being the closest to the camera and – the furthest from the camera.

Before projecting the resulting volume to a plane, we need to ensure that the signal from the occluded points does not interfere with the foreground points. To this end, we perform occlusion reasoning using a differentiable ray tracing formulation, similar to Tulsiani2017drc. We convert the occupancies to ray termination probabilities as follows:

(4) |

Intuitively, a cell has high termination probability if its occupancy value is high and all previous occupancy values are low. The additional background cell serves to ensure that the termination probabilities sum to .

Finally, we project the volume to the plane:

(5) |

Here is the signal being projected, which defines the modality of the result. To obtain a silhouette, we set . For a depth map, we set . Finally, to project a signal associated with the point cloud, such as color, we set to a discretized version of the normalized signal distribution: .

### 4.1 Implementation details

Technically, the most complex part of the algorithm is the conversion of a point cloud to a volume. We have experimented with two implementations of this step: one that is simple and flexible (we refer to it as basic) and another version that is less flexible, but much more efficient (we refer to it as fast

). We implemented both versions using standard Tensorflow

tensorflow2016 operations. At a high level, in the basic implementation each function is computed on an individual volumetric grid, and the results are summed. This allows for flexibility in the choice of the function class, but leads to both computational and memory requirements growing linearly with both the number of points and the volume of the grid , resulting in the complexity . The fast version scales more gracefully, as . This comes at the cost of using the same kernel for all functions . The fastimplementation performs the operation in two steps: first putting all points on the grid with trilinear interpolation, then applying a convolution with the kernel. Further details are provided in Appendix

A.2.## 5 Experiments

### 5.1 Experimental setup

Datasets. We conduct the experiments on 3D models from the ShapeNet (shapenet2015) dataset. We focus on 3 categories typically used in related work: chairs, cars, and airplanes. We follow the train/test protocol and the data generation procedure of Tulsiani2017drc: split the models into training, validation and test sets and render random views of each model with random light source positions and random camera azimuth and elevation, sampled uniformly from and respectively.

Evaluation metrics.

We use the Chamfer distance as our main evaluation metric, since it has been shown to be well correlated with human judgment of shape similarity

(Sun2018). Given a ground truth point cloud and a predicted point cloud , the distance is defined as follows:(6) |

The two sums in Eq. (6) have clear intuitive meanings. The first sum evaluates the precision of the predicted point cloud by computing how far on average is the closest ground truth point from a predicted point. The second sum measures the coverage of the ground truth by the predicted point cloud: how far is on average the closest predicted point from a ground truth point.

For measuring the pose error, we use the same metrics as Tulsiani2018: accuracy (the percentage of samples for which the predicted pose is within of the ground truth) and the median error (in degrees). Before starting the pose and shape evaluation, we align the canonical pose learned by the network with the canonical pose in the dataset, using Iterative Closest Point (ICP) algorithm on the first models in the validation set. Further details are provided in Appendix A.3.

Training details. We trained the networks using the Adam optimizer KingmaBa2015, for mini-batch iterations. We used mini-batches of samples ( views of objects). We used a fixed learning rate of and the standard momentum parameters. We used the fast projection in most experiments, unless mentioned otherwise. We varied both the number of points in the point cloud and the resolution of the volume used in the projection operation depending on the resolution of the ground truth projections used for supervision. We used the volume with the same side as the training samples (e.g., volume for projections), and we used points for projections, points for projections, and points for projections.

When predicting dense point clouds, we have found it useful to apply dropout to the predictions of the network to ensure even distribution of points on the shape. Dropout effects in selecting only a subset of all predicted points for projection and loss computation. In experiments reported in Sections 5.2 and 5.3 we started with a very high dropout and linearly reduced it to towards the end of training. We also implemented a schedule for the point size parameters, linearly decreasing from of the projection volume size to over the course of training. The scaling coefficient of the points was learned in all experiments. An ablation study is shown in Appendix B.1.

Computational efficiency. A practical advantage of a point-cloud-based method is that it does not require using a 3D convolutional decoder as required by voxel-based methods. This improves the efficiency and allows the method to better scale to higher resolution. For resolution the training times of the methods are roughly on par. For the training time of our method is roughly day in contrast with days for its voxel-based counterpart. For the training time of our method is days, while the voxel-based method does not fit into Gb of GPU memory with our batch size.

Resolution | Resolution | Resolution | ||||||
---|---|---|---|---|---|---|---|---|

DRC (Tulsiani2017drc) | PTN (Yan2016) | Ours-V | Ours | Ours-V | Ours | EPCG (Lin2018) | Ours | |

Airplane | 4.03 | |||||||

Car | 3.69 | |||||||

Chair | 5.62 | |||||||

Mean | 4.45 |

### 5.2 Estimating shape with known pose

Comparison with baselines.

We start by benchmarking the proposed formulation against existing methods in the simple setup with known ground truth camera poses and silhouette-based training. We compare to Perspective Transformer Networks (PTN) of

Yan2016, Differentiable Ray Consistency (DRC) of Tulsiani2017drc, Efficient Point Cloud Generation (EPCG) of Lin2018, and to the voxel-based counterpart of our method. PTN and DRC are only available for output voxel grid resolution. EPCG uses the point cloud representation, same as our method. However, in the original work EPCG has only been evaluated in the unrealistic setup of having 100 random views per object and pre-training from 8 fixed views (corners of a cube). We re-train this method in the more realistic setting used in this work – 5 random views per object.Input | View 1 | View 2 | Input | View 1 | View 2 | Input | View 1 | View 2 |
---|---|---|---|---|---|---|---|---|

The quantitative results are shown in Table 1. Our point-cloud-based formulation (Ours) outperforms its voxel-based counterpart (Ours-V) in all cases. It improves when provided with high resolution training signal, and benefits from it more than the voxel-based method. Overall, our best model (at 128 resolution) decreases the mean error by compared to the best baseline. An interesting observation is that at low resolution, PTN performs remarkably well, closely followed by our point-cloud-based formulation. Note, however, that the PTN formulation only applies to learning from silhouettes and cannot be easily generalized to other modalities.

Our model achieves improvement over the point cloud method EPCG, despite it being trained from depth maps, which is a stronger supervision compared to silhouettes used for our models. When trained with silhouette supervision only, EPCG achieves an average error of , times worse than our model. We believe our model is more successful because our rendering procedure is differentiable w.r.t. all three coordinates of points, while the method of Lin et al. – only w.r.t. the depth.

Input |
||||||||
---|---|---|---|---|---|---|---|---|

Ground truth |
||||||||

MVC (Tulsiani2018) |
||||||||

Ours-naive |
||||||||

Ours |

Colored point clouds. Our formulation supports training with other supervision than silhouettes, for instance, color. In Figure 4 we demonstrate qualitative results of learning colored point clouds with our method. Despite challenges presented by the variation in lighting and shading between different views, the method is able to learn correctly colored point clouds. For objects with complex textures the predicted colors get blurred (last example).

Learnable covariance. In the experiments reported above we have learnt point clouds with all points having identical isotropic covariance matrices. We conducted additional experiments where covariance matrices are learnt jointly with point positions, allowing for more flexible representation of shapes. Results are reported in Appendix B.3.

### 5.3 Estimating shape and pose

We now drop the unrealistic assumption of having the ground truth camera pose during training and experiment with predicting both the shape and the camera pose. We use the ground truth at pixel resolution for our method in these experiments. We compare to the concurrent Multi-View Consistency (MVC) approach of Tulsiani2018, using results reported by the authors for pose estimation and pre-trained models provided by the authors for shape evaluations.

Quantitative results are provided in Table 2. Our naive model (Ours-naive) learns quite accurate shape ( worse than MVC), despite not being able to predict the pose well. Our explanation is that predicting wrong pose for similarly looking projections does not significantly hamper the training of the shape predictor. Shape predicted by the full model (Ours) is yet more precise: more accurate than MVC and only less accurate than with ground truth pose (as reported in Table 1). Pose prediction improves dramatically, thanks to the diverse ensemble formulation. As a result, our pose prediction results are on average slightly better than those of MVC (Tulsiani2018) in both metrics, and even better in median error than the results of training with ground truth pose labels (as reported by Tulsiani2018).

Shape () | Pose (Accuracy & Median error) | ||||||||||

MVC (Tulsiani2018) | Ours-naive | Ours | GT pose (Tulsiani2018) | MVC (Tulsiani2018) | Ours-naive | Ours | |||||

Airplane | |||||||||||

Car | |||||||||||

Chair | |||||||||||

Mean |

Figure 5 shows a qualitative comparison of shapes generated with different methods. Even the results of the naive model (Ours-naive) compare favorably to MVC (Tulsiani2018)

. Introducing the pose ensemble leads to learning more accurate pose and, as a consequence, more precise shapes. These results demonstrate the advantage of the point cloud representation over the voxel-based one. Point clouds are especially suitable for representing fine details, such as thin legs of the chairs. (Note that for MVC we use the binarization threshold that led to the best quantitative results.) We also show typical failure cases of the proposed method. One of the airplanes is rotated by

degrees, since the network does not have a way to find which orientation is considered correct. The shapes of two of the chairs somewhat differ from the true shapes. This is because of the complexity of the training problem and, possibly, overfitting. Yet, the shapes look detailed and realistic.### 5.4 Discovery of semantic correspondences

Templates | |||||
---|---|---|---|---|---|

Besides higher shape fidelity, the “matter-centric” point cloud representation has another advantage over the “space-centric” voxel representation: there is a natural correspondence between points in different predicted point clouds. Since we predict points with a fully connected layer, the points generated by the same output unit in different shapes can be expected to carry similar semantic meaning. We empirically verify this hypothesis. We choose two instances from the validation set of the chair category as templates (shown in the top-left corner of Figure 6) and manually annotate 3D keypoint locations corresponding to characteristic parts, such as corners of the seat, tips of the legs, etc. Then, for each keypoint we select all points in the predicted clouds within a small distance from the keypoint and compute the intersection of the points indices between the two templates. (Intersection of indices between two object instances is not strictly necessary, but we found it to slightly improve the quality of the resulting correspondences.) We then visualize points with these indices on several other object instances, highlighting each set of points with a different color. Results are shown in Figure 6. As hypothesized, selected points tend to represent the same object parts in different object instances. Note that no explicit supervision was imposed towards this goal: semantic correspondences emerge automatically. We attribute this to the implicit ability of the model to learn a regular, smooth representation of the output shape space, which is facilitated by reusing the same points for the same object parts.

## 6 Conclusion

We have proposed a method for learning pose and shape of 3D objects given only their 2D projections, using the point cloud representation. Extensive validation has shown that point clouds compare favorably with the voxel-based representation in terms of efficiency and accuracy. Our work opens up multiple avenues for future research. First, our projection method requires an explicit volume to perform occlusion reasoning. We believe this is just an implementation detail, which might be relaxed in the future with a custom rendering procedure. Second, since the method does not require accurate ground truth camera poses, it could be applied to learning from real-world data. Learning from color images or videos would be especially exciting, but it would require explicit reasoning about lighting and shading, as well as dealing with the background. Third, we used a very basic decoder architecture for generating point clouds, and we believe more advanced architectures (Yang2018) could improve both the efficiency and the accuracy of the method. Finally, the fact that the loss is explicitly computed on projections (in contrast with, e.g., Tulsiani2017drc), allows directly applying advanced techniques from the 2D domain, such as perceptual losses and GANs, to learning 3D representations.

#### Acknowledgements

We would like to thank René Ranftl and Stephan Richter for valuable discussions and feedback. We would also like to thank Shubham Tulsiani for providing the models of the MVC method for testing.

## References

## Appendices

## Appendix A Implementation details

### a.1 Network architecture

The convolutional encoder includes layers. The first one has a kernel with channels and stride . The remaining layers all have kernels and come in pairs. The first layer in the pair has stride , the second one – stride . The number of channels grows by a factor of after each strided layer. The convolutional encoder is followed by two fully connected layers with units. Then the network separates into two branches predicting shape and pose. The shape branch has one hidden layer with units and then predicts the point cloud. The pose branch has one shared hidden layer with units. In the naive variant of the method, pose is predicted directly from this hidden layer. In the full approach with an ensemble of pose predictors, this layer is followed by separate hidden layers for each pose predictor in the ensemble, with

units each. We used leaky ReLU with the negative slope

after all layers except for the shape prediction layer where we used the non-linearity to constrain the output coordinates.### a.2 Differentiable point cloud projection

Assume we are given a set of points with coordinates and sizes , as well as the desired spatial dimensions

of the volume to be used for projection. Here we assume indexing of all tensors is 0-based.

In the basic implementation, we start by creating a coordinate tensor of dimensions with entries . Next, for each point we compute the corresponding Gaussian:

(7) |

Finally, we sum these to get the resulting volume: . This implementation is simple and allows for independently changing the sizes of points. However, on the downside, both memory and computation requirements scale linearly with the number of points.

Since linear scaling with the number of points makes large-scale experiments impractical, we implemented the fast version of the method that has lower computation and memory requirements. We implement the conversion procedure as a composition of trilinear interpolation and a convolution. Efficiency comes at the cost of using the same kernel for all points. We implemented trilinear interpolation using the Tensorflow scatter_nd function. We used standard 3D convolutions for the second step. For improved efficiency, we factorized them into three 1D convolutions along the three axes.

### a.3 Quantitative evaluation

To extract a point cloud from the ground truth meshes, we used the vertex densification procedure of Lin2018. For the outputs of voxel-based methods, we extract the surface mesh with the marching cubes algorithm and sample roughly points from the computed surface. We tuned the threshold parameters of the marching cubes algorithm based on the Chamfer distance on the validation set.

For pose evaluations, we computed the angular difference between two rotations represented with quaternions and as .

## Appendix B Additional experiments

### b.1 Ablation study

We evaluate the effect of different components of the model on the shape prediction quality. We measure these by training with pose supervision on ShapeNet chairs, with resolution of the training images. Results are presented in Table 3. The “Full” method is trained with point, point dropout, sigma schedule, and learned point scale. All our techniques are useful, but generally the method is not too sensitive to these.

Full | No drop. | Fixed | Fixed scale | 4000 pts. | 2000 pts. | 1000 pts. | |
---|---|---|---|---|---|---|---|

Precision | |||||||

Coverage | |||||||

Chamfer |

### b.2 Additional qualitative results

Additional qualitative results are shown in Figure 7.

Input |
||||||||
---|---|---|---|---|---|---|---|---|

Ground truth |
||||||||

MVC (Tulsiani2018) |
||||||||

Ours |
||||||||

Input |
||||||||

Ground truth |
||||||||

MVC (Tulsiani2018) |
||||||||

Ours |
||||||||

Input |
||||||||

Ground truth |
||||||||

MVC (Tulsiani2018) |
||||||||

Ours |

### b.3 Towards part-based models

Input |
|||
---|---|---|---|

Isotropic |
|||

Full cov. |

In most experiments in the paper the shape parameters of the points were set by hand, and only the scaling factor was learned. However, our formulation allows learning the shape parameters jointly with the positions of the points. Here we explore this direction using the basic implementation, since it allows for learning a separate shape for each point in the point set. We explore two possibilities: isotropic Gaussians, parametrized by a single scalar and general covariance matrices, parametrized by numbers: diagonal values and a quaternion representing the rotation (this is an overcomplete representation). This resembles part-based models: now instead of composing the object of “atomic” points, a whole object part can be represented by a single Gaussian of appropriate shape (for instance, an elongated Gaussian can represent a leg of a chair).

Figure 8 qualitatively demonstrates the advantage of the more flexible model over the simpler alternative with isotropic Gaussians. One could imagine employing yet more general and flexible per-point shape models, and we see this as an exciting direction of future work.

Figure 9 shows the projection error of different approaches for varying number of points in the set. Learnable parameters perform better than hand-tuned and learned full covariance performs better than learned isotropic covariance. A caveat is that training with full covariance matrix is computationally more heavy in our implementation.

### b.4 Additional visualizations of semantic correspondences

Additional visualizations of semantic correspondences are shown in Figure 10. We use the same two templates here as in the main paper.

### b.5 Interpolation of shapes in the latent space

Fig. 11 shows results of linear interpolation between shapes in the latent space given by the first (shared) fully connected layer. We can observe gradual transitions between shapes, which indicates that the model learns a smooth representation of the shape space. Failure cases, such as legs in the second row, can be attributed to the limited representation of the office chairs with legs in the dataset.

Input A | Prediction A | Prediction B | Input B | |||
---|---|---|---|---|---|---|

Comments

There are no comments yet.