A coarse-to-fine algorithm for registration in 3D street-view cross-source point clouds

10/24/2016 ∙ by Xiaoshui Huang, et al. ∙ Tsinghua University Nokia 0

With the development of numerous 3D sensing technologies, object registration on cross-source point cloud has aroused researchers' interests. When the point clouds are captured from different kinds of sensors, there are large and different kinds of variations. In this study, we address an even more challenging case in which the differently-source point clouds are acquired from a real street view. One is produced directly by the LiDAR system and the other is generated by using VSFM software on image sequence captured from RGB cameras. When it confronts to large scale point clouds, previous methods mostly focus on point-to-point level registration, and the methods have many limitations.The reason is that the least mean error strategy shows poor ability in registering large variable cross-source point clouds. In this paper, different from previous ICP-based methods, and from a statistic view, we propose a effective coarse-to-fine algorithm to detect and register a small scale SFM point cloud in a large scale Lidar point cloud. Seen from the experimental results, the model can successfully run on LiDAR and SFM point clouds, hence it can make a contribution to many applications, such as robotics and smart city development.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 4

page 5

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Researchers have shown great interests in the applications of registration, such as view searching in smart city, location-based service, street view reconstruction and augmented reality. With the progress of sensing technology, many types of 3D point cloud sensor have been developed. Coming from different types of sensors, point clouds are cross-source ones. Compared to same source, registration on cross-source point clouds show great generalization. In this paper, we propose a method to conduct the registration of a small-scale SFM point cloud on a large-scale street-view point cloud.

Figure 1 is a typical example of cross-source point clouds, which contains cross-source problems. At least four challenges can be posed in solving cross-source point cloud registration. (1) Density variation. Different sampling density and sampling theory from various sensors result in distinguished point number in two types of point clouds. Therefore, one point cloud may be much denser than the others. (2) Scale variation. Due to their different sampling density, the scale is hard to maintain same metric in two types of sensor. Furthermore, in terms of the point clouds reconstructed by structure from motion software (e.g. VSFM[vsfm]) or SLAM [huang2016real]

, the scale information is usually unknown, thus we need to estimate the scale. (3) Noise, outliers and missing data. Different sensing mechanisms make a large amount of noise and outliers in the cross-source point clouds; and some parts of the scenes cannot be produced points in point clouds. For example, VSFM is unable to generate points in uniform image. (4) Viewpoint variation. Viewpoint divergence, which is normal in cross-source point clouds, makes the describing content originally different to some extent.

Fig. 1: An example of cross-source point clouds of SFM and LiDAR highlighted from the street view scene. The top left is the SFM point cloud and the top right is the detected registration result on LiDAR point cloud.

With respect to point cloud registration, the existing methods can be casted into two categories: direct and transformation methods. Among them, direct methods directly use the coordinate of point clouds and extract some descriptors to assistant matching and registration, with typical examples of ICP [icp] and other feature-based methods [hontani2010point, sharp2002icp, huang2015dense, huang2015graph] . These methods show great ability in same source point clouds while different extent of limitation on cross-source point clouds. To be specific, they are aimed at looking for an exact matching point for each point, which takes up only a very small proportion in cross-soured point clouds though. Consequently, direct methods make matcher point searching confused all the time when there is a large amount of outliers and noises. In addition, they rely on initialization. For example, similar to our work, Peng. et al [7025406]

proposed two-stage algorithm by using ICP. Due to the ICP’s limitation, the final registration results accuracy is low, which is visually not registered correctly. In terms of transformation methods, they firstly transform 3D point clouds into other space or other model, and then use these transformed data for matching and registration, with typical example of Gaussian mixture model (GMM)

[JR-MPC, jian2011robust], using GMM to describe point cloud and matching them. These methods focus more on global information while ignore local structure distortion. It shows many advantages than direct methods in dealing with cross-source problems. Our algorithm belongs to transformation methods.

In this paper, a novel coarse-to-fine algorithm is proposed to register two cross-source point clouds (one is whole street, the other is a small part) . There are mainly two steps: 1) top K potential regions are detected by a coarse matching in the large-scale LiDAR street-view point cloud for a small-scale SFM point cloud. 2) a generative GMM registration method is applied to refine the matching results from the first stage.

The main contribution is a effective coarse-to-fine pipeline which utilizes the concept of GMM to do registration in large-scale cross-source point clouds. Different to previous generative GMM, we propose a coarse-to-fine, scale normalization and complexity reduction strategies to extend it to be suitable to the new cross-source problem. Different to previous ICP-based method [7025406], we utilize the statistic property of cross-source point cloud to deal with the large variations in local points. It successfully overcomes the point-to-point limitation in ICP-based methods.

Ii Coarse-to-fine Algorithm

The algorithm is illustrated in Figure 2. It contains coarse matching and fine registration, among which coarse matching aims at finding the top potential regions in LiDAR point cloud that potentially match with SFM point cloud. It substantially reduces the number of candidate regions and hence saves computation cost of the next stage. We compute ESF ( Ensemble of Shape Functions) descriptors [ESF] of these potential regions and use them to conduct the first coarse matching. Then, a improved generative GMM-based registration is performed to obtain the transformation of two cross-source point clouds and use the transformation error to refine the matching results. Furthermore, main steps of the second stage are (1) obtain transformation matrix of each registration; (2) acquire residual error of each registration by applying transformation matrix to the original two cross-source point clouds (e.g. selected LiDAR region and SFM); (3) use residual error to re-rank the matching results and output the ranked registration results. After registration, accurate transformation matrix is obtained, and they can be used for applications such as location based service.

Fig. 2: Overview of the proposed algorithm.

Ii-a Coarse Matching

This part is aimed at computing Top K potential regions from LiDAR point cloud for the SFM point cloud. We compute ESF descriptor for point clouds regions (LiDAR and SFM point clouds’ region). ESF is a statistical descriptor depicting a point cloud globally. As a 640-bin histogram, it describes the point cloud based on ten different kinds of shape distribution, including three kinds of line distance distribution, three kinds of angle distribution, three kinds of triangle area distribution and one kind of ratio of line distance distribution [ESF]. Each shape distribution is sampled by a 64-bin histogram and all the ten histograms are concatenated to form the ESF descriptor. The ESF is robust to local structure distortion, which is a valuable advantage in describing high noisy cross-source point clouds (street view point clouds are employed in our experiments).

Since scale difference is a common variation in cross-source point clouds, we should remove it before ESF computation. As to matching and registration, our goal is to find a region with largest similarity, thus the matched point clouds should have a large proportion of overlapping. In this paper, we assume that the matched point clouds have the same voxel box, and then the scale can be computed automatically by

(1)

where is the radius of voxel box containing LiDAR point cloud while is the radius of voxel box containing SFM point cloud. For , it can be computed from the point cloud by counting the maximized distance to the center point. For , in order to cover the potential regions of SFM point clouds, multi-scale radius is needed (e.g. the LiDAR radius is defined from 10 meter to 1 kilometer). A case of multi-scale radius selection is shown in our experiments.

With the scale factor computed, a scale transformation to all points in SFM point cloud is carried out.

(2)

where represents the SFM point cloud. After scale normalization, ESF is applied to both LiDAR and SFM point clouds. Then we compute the similarity() between the ESF descriptors by

(3)

where is the ESF descriptor of selected region in LiDAR point cloud, is the ESF descriptor of SFM point cloud. For SFM point cloud, top potential regions in LiDAR point cloud are selected by using these similarity values. In this stage, only top potential regions are selected, however, the accurate registration relations (transformation matrix) is still unknown.

Ii-B Fine Registration

To obtain the correct registration results for SFM point cloud from detected top regions in the large LiDAR point cloud, the fine registration is indispensable. For cross-source point cloud registration, the conventional point-to-point level methods face much difficulty in registering these large variable cross-source point clouds. This is because of their simple least square mean error of point-level correspondence can be easily lead to sub-optimal in the large variant cross-source point clouds. In order to address the limitation issues of previous direct methods in terms of cross-source problems, we use GMM to consider global statistical properties (e.g. global shape or distribution). GMM-based method focuses on whether the two cross-source point clouds are globally registered and ignores the large variation in local structure.

A generative GMM method is presented by considering multiple point clouds from a same GMM named JR-MPC [JR-MPC]. The paper only reports experiments on same source. As the two registered cross-source point clouds are depicting for the same region (e.g. a house or a tree), it is also reasonable to use a GMM to describe the same region and consider these point clouds as two samples for this GMM. Hence, in this paper, we extend this method to cross-source point cloud registration problem. In the following part, we will describe the proposed strategies to extend JR-MPC to cross-source registration problem.

Due to the cross-source problems (discussed in section 1), the above JR-MPC cannot apply to cross-source point cloud registration problems directly. There are two main problems: 1) JR-MPC are originally designed to address multi point sets registration problem. It has the assumption that the point sets under the same scale. 2) In the expectation step of these generative GMM methods, it need to estimate probability of each point belonging to every Gaussian model. The computation and memory complexity are very large which is

, where and are the number of two point sets and is the Gaussian model. In the model of JR-MPC, even worse, the complexity is . If considering , the complexity both in memory and computation is approximately , which is prohibitive for large scale cross-source point cloud. We will describe how to effectively deal with these problems.

Firstly, for the scale problem, it have been normalized in the previous coarse matching stage. Due to the proposed coarse matching stage, it not only provides the top potential regions in the large scene, but also normalize the scale problem that suitable for the registration stage.

Secondly, for the complexity problem, we need to look deeply to the cross-source problem so as to settle the computation and memory complexity problem. Due to large variations in cross-source point clouds, the exact matched points are only a small proportion. So, trying to find exact matched point is a difficult task in changeable data. Instead of using the original changeable points, the statistical property (GMM) shows stable and more value in registering two cross-source point clouds. It avoids the local variations and focus on more on the global mean and variance. Also, in this paper, the cross-source point cloud registration problem only contains rigid transformation. Hence, we uniformly down-sample the point cloud when the points are over two thousand. After down-sampling, the global shape or structure and rigid transformation still keep the same as the original point cloud (see Figure

3). Due to the shape of two point clouds are all kept, the region of GMM depicted are the same. So, the transformation matrix computed by the down-sample point cloud is the same to the original point cloud. In this way, these two point clouds registration can be successfully converted from a large complexity problem to a feasible problem. If a rigid transformation is computed by using down-sampled point cloud, we can directly apply the rigid transformation to the original point clouds and obtain the final registration results.

Fig. 3: Visual results of original and down-sample point clouds.

In the fine registration stage, when the improved JR-MPC is completed, a residual error is computed to re-rank the matching results in the first stage. To compute the residual error, the computed transformation matrix from revised JR-MPC is applied to perform transformation to the original point clouds. Next, the nearest neighbour is computed for each point and the mean residual is computed following. The residual error is computed by

(4)

where is the point in point cloud ; is the nearest neighbor of in the matched point cloud . A lower means the two point clouds are more similar.However, based on our observation, the always shows lower value in small-scale point clouds. To eliminate this scale bias, a penalty is defined related to the scale value:

(5)

where, is the penalty for scale variation. is the parameter to control the penalty, scale is computed by formulation (1). The final ranking regions are sorted by the value and the top ranked one represents the best matching to the SFM point cloud. The pseudo-code of the coarse-to-fine algorithm is shown in Algorithm 1.

0:  
0:  
  
  
  
  
  
  
  
  
  
  
  
Algorithm 1 Pseudocode of coarse-to-fine algorithm

Iii Experimental results

The experiments are conducted on real cross-source point clouds that are combined by LiDAR and SFM point cloud. LiDAR point clouds are captured from three different scenes in Helsinki (Helsinki Cathedral, Helsinki station and Library of University of Helsinki), with hundreds of millions of points on each original LiDAR point cloud. To efficiently match and register on the large volume data, the LiDAR point clouds are down-sampled into 10% of the original points. For SFM point clouds, three typical buildings are selected and 2D images are captured by digit camera. Helsinki Station is divided into two objects: station south and station east.

Fig. 4: Eight point clouds of four objects named Cathedral, Station south, Station east and Library. Each row represents one object with two cross-source point clouds. The left column is LiDAR and the right is SFM.

We use 2D images and VSFM [vsfm] to build a software-reconstructed point clouds. The four objects of LiDAR and SFM point clouds are illustrated in Figure 4. Before applying the proposed algorithm, standard pre-processing, such as removal of sparse outliers, is conducted for both point clouds. Considering computation complexity reduction, the performance of the proposed algorithm is evaluated on a subset data. The subset data is generated by 7 different scale spheres scanning all the LiDAR point clouds. The radius of the spheres ranges from 30 to 60 with an interval of 5. A hundred regions are selected under each scale. The subset data are regarded as candidate regions for matching and registration. The candidate regions will cover more than areas of LiDAR point clouds. The matching and registration is then regarded as a retrieval problem.The target point cloud(SFM) is retrieved from the 700 candidate LiDAR regions (100 candidates for each one of the 7 scales).

Based on our study, the first matching stage can achieve the best performance when the number of ESF sampling level is 64. The number of potential regions kept for the second fine registration stage is selected as 20 in all the experiments.

We define two single stage baseline systems and select ESF+ICP, ESF+GO-ICP as our compared methods. For baseline systems, like [7025406], one is retrieved by ESF only to measure the ESF similarity of point clouds. The other is applying ICP to compute the residual error on each region in every point clouds. Since scale variation is a common problem existing in cross-source point cloud, we normalize the scale by our scale estimation method before applying ICP. In the baseline system, one difference is that the scales do not adjust the residual produced by ICP, as that in the proposed method. To compare the performance of proposed method, we regard the candidate regions which cover area of the target object and points associated with the background are regarded as ground-truth data. In this paper, rank-5 measurement is proposed and the ground-truth number is more than 5. According to their ESF similarity or final residual error, candidate regions selected from LiDAR point clouds are sorted and the rank is cut off at top 5. The algorithm shows better performance when there are more retrieved ground-truth regions. All experiments are conducted in a computer with 4-core 3.2GHz CPU and 8GB memory. The results are illustrated in Table I.

Fig. 5: The top 2 registration results of the 4 objects with the proposed method and [7025406]. Each row represents the results for one object. Figures in the 1th column and the 3th respectively represent the GMM registration results of retrieved rank 1 and rank 2 candidate regions in LiDAR and SFM point cloud. Figures in the 2th to 4th columns respectively represent the registration results of whole LiDAR street scene and SFM point cloud. The target in LiDAR is in red, the reference in yellow and other points are in green. Figure in the 5th column is the top 1 registration result of [7025406], with LiDAR in green and SFM in red.
cathedral library station south station east
accuracy time(s) accuracy time(s) accuracy time(s) accuracy time(s)
Baseline: single stage ESF 4 24 0 25 2 23 0 25
Baseline: single stage ICP 5 305 5 241 0 167 5 522
ESF-64 + ICP without adjusting final residual [7025406] 5 85 3 73 0 56 4 139
ESF-64 + ICP with adjusting final residual [7025406] 5 85 3 73 4 56 5 129
ESF-64 + Go-ICP with adjusting final residual [7025406] 5 85 3 73 4 56 5 129
The proposed method 5 300 5 223 5 320 5 256
TABLE I: The performance of the proposed method and the compared methods

As shown in Table I, the single stage ESF performs faster but suffers from low accuracy. The baseline of single stage ICP, however, possesses higher accuracy, but it is the most time consuming method. The compared method [7025406] runs much faster than the baseline of single stage of ICP and the proposed method. It uses ESF to quickly remove many incorrect candidates, and then ICP is applied to refine the result, saving a large amount of time. However, it shows lower accuracy to the proposed method, which can be visually seen from Station South and Cathedral in Figure 5. Using ESF as well, the generative GMM in the proposed method is based on the assumption that two point clouds come from the same object. If the two point clouds have plenty of differences, they are original not registered and it shows high residual error. Also, library and station south results in Table I show that the proposed algorithm is more robust in registration. The proposed method can be robust in detecting the top 5 ground-truth regions from cross-source point clouds (as described before, the ground-truth regions are more than 5). Therefore, the proposed method not only retrieves the correct regions but also registers them more accurately than the compared methods.

In addition, the proposed method is conducted on the whole data set, where multi-scale regions over the whole scene are tested, which means that much more negative regions are included. The top two registered point clouds of each object are presented in Figure 5. It shows the high registration accuracy of the proposed method from the viewpoint, especially on Station South and Cathedral.

Iv Conclusion

In this paper, a novel coarse to fine algorithm is proposed to address the problem of cross-source point cloud registration. In the fist stage, coarse matching is performed to quickly detect a few potential matched regions. In the second stage, two revisions about scale and complexity are proposed to extend the recent generative Gaussian mixture model method to cross-source point cloud registration problem. It refines the matching results and accurately finds out registration regions. The proposed method does not rely on least square mean error, but rather utilizes the statistical property that is robust to the cross-source problems. It can efficiently detect the potential regions from a large scene and register them accurately. The proposed method can not only detect where the reference point cloud is located in the big scene but also obtain the accurate pose related to the big scene. The future work is to develop many applications with this method in areas such as location-based service in smart city and robotics.

References