Voxel-based Network for Shape Completion by Leveraging ...

14
Voxel-based Network for Shape Completion by Leveraging Edge Generation - Supplementary Material Xiaogang Wang Marcelo H Ang Jr Gim Hee Lee National University of Singapore [email protected] {mpeangh,gimhee.lee}@nus.edu.sg 1. Ground Truth Edge Generation We obtain the ground truth edges b P e with [1]. Specif- ically, object edges are identified by evaluating the query point p from its k-nearest neighbors. We first find the k- nearest neighbors of each query point from the object and denote c as the center of these neighbors. We then calculate the minimum distance v among all the neighboring points to the query point. A query point p is classified as the edge point if ||c p|| · v. We set λ =5 and k = 100 for our created dataset, and set λ =1.8 and k = 150 for the Completion3D dataset. 2. Evaluation Metrics Following previous methods [10, 6, 7, 9], we use the CD and Fr´ echet Point Cloud Distance (FPD) [7, 5] as the evalu- ation metrics for the synthetic datasets and fidelity and reg- istration errors for the KITTI dataset. FPD. FPD evaluates the distribution similarity by the 2- Wasserstein distance between the real and fake Gaussian measured in the feature spaces of the point sets, i.e. FPD(X, Y)= m X m Y 2 2 + TrX Y 2(Σ X Σ Y ) 1 2 ). (1) Fidelity. Fidelity measures the average distance from each point in the input to its nearest neighbor in the out- put. It evaluates how well the input points are preserved in the output. Registration Errors. Rotation and translation errors are the evaluation metrics for the point cloud registration. More specially, it measures the registration performances between neighboring frames in the same Velodyne sequence. Two types of inputs are evaluated: the partial points from the raw scans, and the generated complete points by different models. The rotation error is computed as 2cos 1 (2 < q 1 ,q 2 > 2 1), where q 1 and q 2 are the ground truth rota- tion and the rotation computed from ICP, respectively. The translation error is computed as ||t 1 t 2 || 2 , in which t 1 is the ground truth translation and t 2 is the translation gener- ated by ICP, respectively. 3. More Details of Conversion between Points and Grids 1) Conversion from points to grids. We calculate the initial grid features as the coordinate differences between points and their corresponding eight nearby grid vertexes in five different scales. This results in five tensors of sizes 3 × 2048 × 8, 3 × 1024 × 8, 3 × 512 × 8, 3 × 256 × 8 and 3 × 128 × 8, respectively. Corresponding to the differ- ent point resolutions {2048, 1024, 512, 256, 128}, the five voxel resolutions are {32 3 , 16 3 , 8 3 , 4 3 , 2 3 }. A set of grid features P i f , {i =0, 1, 2, 3, 4} are obtained from the initial grid features by several convolutional blocks (§3.2). The quantitative comparison between our proposed grid trans- formation and the Gridding of GRNet is shown in rows 3 and 4 of Table 6 in the main paper, i.e., our transformation achieves 24.5% relative improvement compared to Grid- ding (3.600 vs 4.768). 2) Conversion from grids to points. GRNet predicts 262,144 (64 3 ) points from every vertex feature, and then samples 2048 points as the coarse output P C and use MLPs to generate dense points from P C . In contrast, we directly predict the dense points by adding the point offsets to the grid centers. The number of points for each grid cell are de- cided by the binary score p c and density value δ c (§3.3.3). 4. Network Architecture Details We express the 3D convolutions with its number of out- put channels, the kernel size, the stride and padding val- ues. For example, C3D(O1K3S1P1) indicates a 3D convo- lutional layer with the number of output channel as 1, kernel size as 3 × 3 × 3, the stride and padding values as 1. DC3D

Transcript of Voxel-based Network for Shape Completion by Leveraging ...

Page 1: Voxel-based Network for Shape Completion by Leveraging ...

Voxel-based Network for Shape Completion by Leveraging Edge Generation -Supplementary Material

Xiaogang Wang Marcelo H Ang Jr Gim Hee LeeNational University of Singapore

[email protected] {mpeangh,gimhee.lee}@nus.edu.sg

1. Ground Truth Edge Generation

We obtain the ground truth edges Pe with [1]. Specif-ically, object edges are identified by evaluating the querypoint p from its k-nearest neighbors. We first find the k-nearest neighbors of each query point from the object anddenote c as the center of these neighbors. We then calculatethe minimum distance v among all the neighboring pointsto the query point. A query point p is classified as the edgepoint if ||c − p|| > λ · v. We set λ = 5 and k = 100 forour created dataset, and set λ = 1.8 and k = 150 for theCompletion3D dataset.

2. Evaluation Metrics

Following previous methods [10, 6, 7, 9], we use the CDand Frechet Point Cloud Distance (FPD) [7, 5] as the evalu-ation metrics for the synthetic datasets and fidelity and reg-istration errors for the KITTI dataset.

FPD. FPD evaluates the distribution similarity by the 2-Wasserstein distance between the real and fake Gaussianmeasured in the feature spaces of the point sets, i.e.

FPD(X,Y) = ∥mX − mY∥22 + Tr(ΣX +ΣY − 2(ΣXΣY)12 ).

(1)

Fidelity. Fidelity measures the average distance fromeach point in the input to its nearest neighbor in the out-put. It evaluates how well the input points are preserved inthe output.

Registration Errors. Rotation and translation errors arethe evaluation metrics for the point cloud registration. Morespecially, it measures the registration performances betweenneighboring frames in the same Velodyne sequence. Twotypes of inputs are evaluated: the partial points from theraw scans, and the generated complete points by differentmodels. The rotation error is computed as 2cos−1(2 <

q1, q2 >2 −1), where q1 and q2 are the ground truth rota-tion and the rotation computed from ICP, respectively. Thetranslation error is computed as ||t1 − t2||2, in which t1 isthe ground truth translation and t2 is the translation gener-ated by ICP, respectively.

3. More Details of Conversion between Pointsand Grids

1) Conversion from points to grids. We calculate theinitial grid features as the coordinate differences betweenpoints and their corresponding eight nearby grid vertexesin five different scales. This results in five tensors of sizes3 × 2048 × 8, 3 × 1024 × 8, 3 × 512 × 8, 3 × 256 × 8and 3× 128× 8, respectively. Corresponding to the differ-ent point resolutions {2048, 1024, 512, 256, 128}, the fivevoxel resolutions are {323, 163, 83, 43, 23}. A set of gridfeatures Pi

f , {i = 0, 1, 2, 3, 4} are obtained from the initialgrid features by several convolutional blocks (§3.2). Thequantitative comparison between our proposed grid trans-formation and the Gridding of GRNet is shown in rows 3and 4 of Table 6 in the main paper, i.e., our transformationachieves 24.5% relative improvement compared to Grid-ding (3.600 vs 4.768).

2) Conversion from grids to points. GRNet predicts262,144 (643) points from every vertex feature, and thensamples 2048 points as the coarse output PC and use MLPsto generate dense points from PC . In contrast, we directlypredict the dense points by adding the point offsets to thegrid centers. The number of points for each grid cell are de-cided by the binary score pc and density value δc (§3.3.3).

4. Network Architecture Details

We express the 3D convolutions with its number of out-put channels, the kernel size, the stride and padding val-ues. For example, C3D(O1K3S1P1) indicates a 3D convo-lutional layer with the number of output channel as 1, kernelsize as 3× 3× 3, the stride and padding values as 1. DC3D

Page 2: Voxel-based Network for Shape Completion by Leveraging ...

PCN [10] PCN-FC [10] CDA [2] TopNet [6] CRN [7] GRNet [9] DPC [11] MSN [3] VE-PCNPara. (M) 6.85 53.2 51.85 9.96 5.14 76.71 6.66 30.32 35.00Time (ms) 57.5 21.4 614.9 63.1 61.3 124.3 331.3 346.7 450.1

Table 1: Space and time comparisons of different methods.

Methods CDfull pipeline 2.669without LS

CD 2.886without LCD 3.408without Ld 2.843

without LEBCE 2.842

Table 2: Ablation studies on the different losses. Resultsare obtained by evaluating mean CD per point (10−4) onour dataset.

represents a 3D deconvolutional layer. We set the dilationvalue for all our convolutions as 1 except the first 3D con-volution in the residual blocks of the edge generator.

The network architectures of the edge generator andshape encoder are shown in Figures 1 and 2, respectively.Figure 3 shows architectures of the refinement cells and theshape decoder. Figure 3 (a) and (b) show the first four re-finement cells and the last refinement cell, respectively. Ev-ery cell shares similar architectures but with different fea-ture dimensions. C1, O1 and O2 in the first four refinementcells are {128, 128, 128}, {256, 128, 128}, {192, 64, 64}and {129, 32, 32}, respectively.

Figure 3 (c) and (d) show the architecture of point gen-erator in the shape completion module and the edge gener-ator. Figure 3 (c) illustrates the prediction architectures forthe classification score pc and density value δc of each gridcell. Figure 3 (d) presents the point set generation for eachgrid. Prior to feeding the grid features into the convolutionallayers, grid features are concatenated with 2 dimensionalrandomly sampled values to increase the point diversity ina local patch.

5. Time and Space Complexity AnalysisThe number of parameters and inference time of differ-

ent methods are shown in Table 1. Some layers in our net-work are 1D CNNs instead of 3D CNNs (e.g. the architec-ture of point generator) and thus we consume much smallerparameters than voxel-based methods (CDA and GRNet).We compute the average inference time of 5000 forwardsteps on a Titan X GPU, and our time complexity is compa-rable to other methods.

6. More Ablation StudiesMore ablation studies on different losses are shown in

Table 2. We test the effects of LSCD, LCD, Lo and LE

BCE by

setting the corresponding weights to be 0. The results areobtained by testing on our created dataset.

7. More Experimental Results7.1. Results on the PCN Dataset

We show the results of our method on the PCN dataset inTable 3. All the other results are cited from the state-of-the-art work PMP-Net [8]. We achieve lower average CD errorscompared to all prior works and obtain better performanceson the majority of object categories.

7.2. More Results on the Completion3D Dataset

More qualitative results on the validation data are shownin Figures 4 and 5.

7.3. More Results on our Dataset

Table 4 shows the FPD evaluations from various meth-ods. More qualitative results on seen categories of ourdataset are shown in Figures 6 and 7. More qualitative re-sults on unseen categories are shown in Figure 8.

To further test the robustness of different models, weocclude the partial input with different occlusion ratios pthat ranges from 20% to 40%. We directly adopt the mod-els trained on seen categories for testing. The quantitativeresults are shown in Table 5. More qualitative results areshown in Figures 9 and 10.

7.4. More Results on the KITTI Dataset

Figure 11 shows the completion results on the KITTIdataset. We evaluate the performances by calculating theregistration errors following PCN [10].

7.5. More Results on the Edge Generation

We show more experimental results on point cloud edgegeneration in Figure 12.

8. Comparisons to SK-PCN [4]SK-PCN proposes a similar thought that adopts skele-

ton generations to help the shape completion. However, ouredges are different from their meso-skeleton. The differ-ences are shown in Figure 13 (The results of SK-PCN areFig. 9 of their paper). Their meso-skeleton focus on theoverall shapes. In contrast, our edges focus on high fre-quency components (e.g. thin structures), which are dif-ficult to generate in existing methods. This can be evi-denced from the limitation cases in Fig. 10 of SK-PCN

Page 3: Voxel-based Network for Shape Completion by Leveraging ...

MethodsMean Chamfer Distance (CD) per point (10−3)

Average Plane Cabinet Car Chair Lamp Sofa Table Vessel

FoldingNet 14.31 9.49 15.80 12.61 15.55 16.41 15.97 13.65 14.99TopNet 12.15 7.61 13.31 10.90 13.82 14.44 14.78 11.22 11.12

AtlasNet 10.85 6.37 11.94 10.10 12.06 12.37 12.99 10.33 10.61PCN 9.64 5.50 10.63 8.70 11.00 11.34 11.68 8.59 9.67

GRNet 8.83 6.45 10.37 9.45 9.41 7.96 10.51 8.44 8.04CRN 8.51 4.79 9.97 8.31 9.49 8.94 10.69 7.81 8.05

PMP-Net 8.66 5.50 11.10 9.62 9.47 6.89 10.74 8.77 7.19Ours 8.32 4.80 9.85 9.26 8.90 8.68 9.83 7.30 7.93

Table 3: Quantitative results on the PCN dataset.

PCN [10] PCN-FC [10] CDA [2] TopNet [6] CRN [7] GRNet [9] DPC [11] MSN [3] VE-PCNFPD 5.584 6.634 9.142 7.7536 3.054 6.513 8.347 3.904 1.882

Table 4: FPD comparisons on different methods. The lower, the better.

RatiosMean Chamfer Distance per point (10−4)

PCN [10] PCN-FC [10] CDA [2] TopNet [6] CRN [7] GRNet [9] DPC [11] MSN [3] VE-PCN

20% 5.642 6.230 7.485 6.732 3.554 4.016 2.975 3.213 2.56530% 5.991 6.704 7.771 7.245 3.932 4.646 4.681 4.084 3.05540% 7.066 7.934 8.741 8.622 5.094 10.746 8.210 6.546 5.129

Table 5: Quantitative comparison of occluded point clouds under different occlusion rates.

paper. Moreover, SK-PCN generates the complete pointsby learning displacements from skeletal points with a localadjustment strategy. In contrast, we synthesize the completepoints by injecting the edge features into the completiondecoder with a voxelization strategy. This voxel structurefurther enables our point generation to be confined withinwell-defined spaces of the grid cells, and thus eliminatesthe generation of spurious points that are commonly seen inother methods.

Page 4: Voxel-based Network for Shape Completion by Leveraging ...

Inst

ance

Nor

m3D

(64)

ReL

U

128X

43 Con

v3D

Inst

ance

Nor

m3D

ReL

U

128X

33 Con

v3D

Inst

ance

Nor

m3D

ReL

U

128X

33 Con

v3D

Inst

ance

Nor

m3D

128X

33 Con

v3D

Inst

ance

Nor

m3D

ReL

U

128X

33 Con

v3D

Inst

ance

Nor

m3D

128X

33 Con

v3D

Inst

ance

Nor

m3D

ReL

U

128X

33 Con

v3D

Inst

ance

Nor

m3D

64X

43 Dco

nv3D

Inst

ance

Nor

m3D

ReL

U

32X

43 Dco

nv3D

Inst

ance

Nor

m3D

ReL

U

1X33 C

onv3

D

1X33 C

onv3

D

C3D

(O64

K4S

2P1)

Inst

ance

Nor

m3D

(128

)

ReL

U

C3D

(O12

8K4S

2P1)

Inst

ance

Nor

m3D

(128

)

ReL

U

C3D

(O12

8K3S

1P2)

Inst

ance

Nor

m3D

(128

)

C3D

(O12

8K3S

1P1)

Inst

ance

Nor

m3D

(128

)

ReL

U

C3D

(O12

8K3S

1P2)

Inst

ance

Nor

m3D

(128

)

C3D

(O12

8K3S

1P1)

Inst

ance

Nor

m3D

(128

)

ReL

U

C3D

(O12

8K3S

1P2)

Inst

ance

Nor

m3D

(128

)

C3D

(O12

8K3S

1P1)

Inst

ance

Nor

m3D

(64)

ReL

U

DC

3D(O

64K

4S2P

1)

C3D

(O1K

3S1P

1)

Inst

ance

Nor

m3D

(32)

ReL

U

DC

3D(O

32K

4S2P

1)

C3D

(O1K

3S1P

1)

Figure 1: Architecture details of the edge generator.

C3D

(O64

K3S

1P1)

Bat

chN

orm

3D(6

4)

ReL

U

Max

Poo

l3D

(K2S

2)

128X

33 Con

v3D

Bat

chN

orm

3D

ReL

U

128X

33 Con

v3D

IBat

chN

orm

3D

ReL

U

Max

Poo

l3D

256X

33 Con

v3D

Bat

chN

orm

3D

ReL

U

256X

33 Con

v3D

IBat

chN

orm

3D

ReL

U

Max

Poo

l3D

512X

33 Con

v3D

Bat

chN

orm

3D

ReL

U

512X

33 Con

v3D

IBat

chN

orm

3D

ReL

U

Max

Poo

l3D

512X

33 Con

v3D

Bat

chN

orm

3D

ReL

U

512X

33 Con

v3D

IBat

chN

orm

3D

ReL

U

Res

hape

C3D

(O64

K3S

1P1)

Bat

chN

orm

3D(6

4)

ReL

U

C3D

(O12

8K3S

1P1)

Bat

chN

orm

3D(1

28)

ReL

U

Max

Poo

l3D

(K2S

2)

C3D

(O12

8K3S

1P1)

Bat

chN

orm

3D(1

28)

ReL

U

C3D

(O25

6K3S

1P1)

Bat

chN

orm

3D(2

56)

ReL

U

Max

Poo

l3D

(K2S

2)

C3D

(O25

6K3S

1P1)

Bat

chN

orm

3D(2

56)

ReL

U

C3D

(O51

2K3S

1P1)

Bat

chN

orm

3D(5

12)

ReL

U

Max

Poo

l3D

(K2S

2)

C3D

(O51

2K3S

1P1)

Bat

chN

orm

3D51

2)

ReL

U

C3D

(O51

2K3S

1P1)

Bat

chN

orm

3D(5

12)

ReL

U

C3D

(O10

24K

2S1P

0)

Bat

chN

orm

3D(1

024)

ReL

U

Res

hape

C3D

(O64

K3S

1P1)

Bat

chN

orm

3D(6

4)

ReL

U

Figure 2: Architecture details of the shape encoder.

Inst

ance

Nor

m3D

Dro

pout

3D

Con

v3D

Inst

ance

Nor

m3D

ReL

U

Dro

pout

3D

Ups

ampl

e3D

Con

v3D

(62+2)X323

64X

1 C

onv1

D

ReL

U

62X323

(62+2)X13

64X

1 C

onv1

D

ReL

U

32X

1 C

onv1

D

ReL

U

32X

1 C

onv1

D

ReL

U

16X

1 C

onv1

D

ReL

U

16X

1 C

onv1

D

ReL

U

8X1

Con

v1D

ReL

U

3X1

Con

v1D

(a)

(d)

62X323

16X

13 Con

v3D

Bat

chN

orm

3D

ReL

U

8X13 C

onv3

D

IBat

chN

orm

3D

ReL

U

4X13 C

onv3

D

Bat

chN

orm

3D

ReL

U

2X13 C

onv3

D

(c)

C3D

(O16

K1S

1P1)

Bat

chN

orm

3D(1

6)

ReL

U

C3D

(O8K

1S1P

1)

Bat

chN

orm

3D(8

)

ReL

U

C3D

(O4K

1S1P

1)

Bat

chN

orm

3D(4

)

ReL

U

C3D

(O2K

1S1P

1)

C1D

(O64

K1S

1P1)

ReL

U

C1D

(O64

K1S

1P1)

ReL

U

C1D

(O32

K1S

1P1)

ReL

U

C1D

(O32

K1S

1P1)

ReL

U

C1D

(O16

K1S

1P1)

ReL

U

C1D

(O16

K1S

1P1)

ReL

U

C1D

(O8K

1S1P

1)

ReL

U

C1D

(O3K

1S1P

1)

C3D

(O2K

3S1P

1)

Inst

ance

Nor

m3D

(C1)

ReL

U

Dro

pout

3D

C3D

(O1K

3S1P

1)

Inst

ance

Nor

m3D

(O1)

Dro

pout

3D

Ups

ampl

e3D

ReL

U

Inst

ance

Nor

m3D

(65)

Dro

pout

3D

C3D

(O62

K3S

1P1)

Inst

ance

Nor

m3D

(62)

ReL

U

(b)

Figure 3: Architecture details of the refinement cells and shape decoder.

Page 5: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN TopNet GRNet VE-PCN GT

Figure 4: Qualitative comparisons on the Completion3D dataset (1/2).

Page 6: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN TopNet GRNet VE-PCN GT

Figure 5: Qualitative comparisons the Completion3D dataset (2/2).

Page 7: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPC MSN26,185,198,208, 275, 297, 333,335,338, 346,389,487,548,

593, 594, 711, 831, 1067, 1094, 1097, 1120, 1124,1128, 1142, 1235,2202

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPCECG MSN

Figure 6: Qualitative comparisons on seen categories of our dataset (1/2).

Page 8: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPC MSN Input PCN CDA TopNet VE-PCN GTCRN GRNet DPCECG MSN

Figure 7: Qualitative comparisons on seen categories of our dataset (2/2).

Page 9: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPC MSN6,30,36,42,60,78,114,180,444,450, Input PCN CDA TopNet VE-PCN GTCRN GRNet DPCECG MSN

Figure 8: Qualitative comparisons on unseen categories of our dataset.

Page 10: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPC MSN72,132,318,336,354,

20%

30%

40%

20%

30%

40%

20%

30%

40%

20%

30%

40%

20%

30%

40%

Input PCN CDA TopNet VE-PCNCRN GRNet DPCECG MSN

Figure 9: Qualitative comparisons on different occlusion ratios (1/2).

Page 11: Voxel-based Network for Shape Completion by Leveraging ...

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPC MSN20%

30%

40%

20%

30%

40%

20%

30%

40%

20%

30%

40%

20%

30%

40%

360,372,894,1116,1758

Input PCN CDA TopNet VE-PCN GTCRN GRNet DPCECG MSN

Figure 10: Qualitative comparisons on different occlusion ratios (2/2).

Page 12: Voxel-based Network for Shape Completion by Leveraging ...

148,

Figure 11: Completion results on KITTI.

Input PCN VE-PCN GT

4.738 3.379

382,

5.924 4.166

3.353 2.5167.614 4.872

7.679 3.162

8.427 5.721

8.847 4.916

8.444 6.321

11.365 6.917

4.886 2.822

8.754 8.284

9.480 4.277

5.943 3.110

Input PCN EG-PCN GT

Figure 12: Edge generation results. Bottom values are theCD errors per point (10−4).

Page 13: Voxel-based Network for Shape Completion by Leveraging ...

SK-PCN

Ours

Input Skeleton GT Input Skeleton GT Input Skeleton GT

Input Edges GT Input Edges GT Input Edges GT

Figure 13: Edge (ours) vs. mes-skeleton (SK-PCN) generation.

Page 14: Voxel-based Network for Shape Completion by Leveraging ...

References[1] Syeda Mariam Ahmed, Yan Zhi Tan, Chee Meng Chew, Ab-

dullah Al Mamun, and Fook Seng Wong. Edge and cornerdetection for unorganized 3d point clouds with applicationto robotic welding. In 2018 IEEE/RSJ International Confer-ence on Intelligent Robots and Systems (IROS), pages 7350–7355. IEEE, 2018.

[2] Isaak Lim, Moritz Ibing, and Leif Kobbelt. A convolutionaldecoder for point clouds using adaptive instance normaliza-tion. In Computer Graphics Forum, volume 38, pages 99–108. Wiley Online Library, 2019.

[3] Minghua Liu, Lu Sheng, Sheng Yang, Jing Shao, and Shi-Min Hu. Morphing and sampling network for dense pointcloud completion. In Proceedings of the AAAI Conferenceon Artificial Intelligence, volume 34, pages 11596–11603,2020.

[4] Yinyu Nie, Yiqun Lin, Xiaoguang Han, Shihui Guo, JianChang, Shuguang Cui, and Jian.J Zhang. Skeleton-bridgedpoint completion: From global inference to local adjustment.In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, andH. Lin, editors, Advances in Neural Information Process-ing Systems, volume 33, pages 16119–16130. Curran Asso-ciates, Inc., 2020.

[5] Dong Wook Shu, Sung Woo Park, and Junseok Kwon. 3dpoint cloud generative adversarial network based on treestructured graph convolutions. In The IEEE InternationalConference on Computer Vision (ICCV), October 2019.

[6] Lyne P Tchapmi, Vineet Kosaraju, S. Hamid Rezatofighi, IanReid, and Silvio Savarese. Topnet: Structural point clouddecoder. In The IEEE Conference on Computer Vision andPattern Recognition (CVPR), 2019.

[7] Xiaogang Wang, Marcelo H. Ang Jr. , and Gim Hee Lee.Cascaded refinement network for point cloud completion.In IEEE/CVF Conference on Computer Vision and PatternRecognition (CVPR), June 2020.

[8] Xin Wen, Peng Xiang, Zhizhong Han, Yan-Pei Cao, PengfeiWan, Wen Zheng, and Yu-Shen Liu. Pmp-net: Point cloudcompletion by learning multi-step point moving paths. InProceedings of the IEEE/CVF Conference on Computer Vi-sion and Pattern Recognition, pages 7443–7452, 2021.

[9] Haozhe Xie, Hongxun Yao, Shangchen Zhou, Jiageng Mao,Shengping Zhang, and Wenxiu Sun. Grnet: Gridding resid-ual network for dense point cloud completion. In EuropeanConference on Computer Vision, pages 365–381. Springer,2020.

[10] Wentao Yuan, Tejas Khot, David Held, Christoph Mertz, andMartial Hebert. Pcn: Point completion network. In 2018 In-ternational Conference on 3D Vision, pages 728–737. IEEE,2018.

[11] Wenxiao Zhang, Qingan Yan, and Chunxia Xiao. Detail pre-served point cloud completion via separated feature aggre-gation. In Computer Vision–ECCV 2020: 16th EuropeanConference, Glasgow, UK, August 23–28, 2020, Proceed-ings, Part XXV 16, pages 512–528. Springer, 2020.