10
Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from Multiple Angles by Joint Self-Reconstruction and Half-to-Half Prediction Zhizhong Han 1,3 , Xiyang Wang 1,2 , Yu-Shen Liu 1,2* , Matthias Zwicker 3 1 School of Software, Tsinghua University, Beijing, China 2 Beijing National Research Center for Information Science and Technology (BNRist) 3 Department of Computer Science, University of Maryland, College Park, USA [email protected], [email protected], [email protected], [email protected] Abstract Unsupervised feature learning for point clouds has been vital for large-scale point cloud understanding. Recen- t deep learning based methods depend on learning global geometry from self-reconstruction. However, these methods are still suffering from ineffective learning of local geome- try, which significantly limits the discriminability of learned features. To resolve this issue, we propose MAP-VAE to enable the learning of global and local geometry by joint- ly leveraging global and local self-supervision. To enable effective local self-supervision, we introduce multi-angle analysis for point clouds. In a multi-angle scenario, we first split a point cloud into a front half and a back half from each angle, and then, train MAP-VAE to learn to pre- dict a back half sequence from the corresponding front half sequence. MAP-VAE performs this half-to-half prediction using RNN to simultaneously learn each local geometry and the spatial relationship among them. In addition, MAP-VAE also learns global geometry via self-reconstruction, where we employ a variational constraint to facilitate novel shape generation. The outperforming results in four shape analy- sis tasks show that MAP-VAE can learn more discriminative global or local features than the state-of-the-art methods. 1. Introduction Point clouds have become a popular 3D representation in machine vision, autonomous driving, and augmented reali- ty, because they are easy to acquire and manipulate. There- fore, point cloud analysis has emerged as a crucial prob- lem in the area of 3D shape understanding. With the help of extensive supervised information, recent deep learning * Corresponding Author. This work was supported by National Key R&D Program of China (2018YFB0505400) and NSF (award 1813583). based feature learning techniques have achieved unprece- dented results in classification, detection and segmenta- tion [27, 31, 22, 35, 23, 39]. However, supervised learning requires intense manual labeling effort to obtain supervised information. Therefore, unsupervised feature learning is an attractive alternative and a promising research challenge. Several studies have tried to adress this challenge [1, 21, 4, 34, 33, 42, 32]. To learn the structure of a point cloud without additional supervision, these generative models are trained by self-supervision, such as self-reconstruction [1, 4, 6, 42, 33, 32] or distribution approximation [1, 21, 34], which is implemented by auto-encoder or generative ad- versarial networks [8] respectively. To capture finer glob- al structure, some methods [34, 33, 42, 32] first learn local structure information in point cloud patches based on which the global point cloud is then reconstructed. Because of lacking effective and semantic local structure supervision, however, error may accumulate in the local structure learn- ing process, which limits the network’s ability in 3D point cloud understanding. To resolve this issue, we propose a novel deep learn- ing model for unsupervised point cloud feature learning by simultaneously employing effective local and global self- supervision. We introduce multi-angle analysis for point clouds to mine effective local self-supervision, and com- bine it with global self-supervision under a variational con- straint. Hence we call our model Multi-Angle Point Cloud Variational Auto-Encoder(MAP-VAE). Specifically, to em- ploy local self-supervision, MAP-VAE first splits a point cloud into a front half and a back half under each of sev- eral incrementally varying angles. Then, MAP-VAE per- forms half-to-half prediction to infer a sequence of several back halves from the corresponding sequence of the com- plementary front halves. Half-to-half prediction aims to capture the geometric and structural information of local regions on the point cloud through varying angles. More-

Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

  • Upload
    others

  • View
    7

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D PointClouds from Multiple Angles by Joint Self-Reconstruction and Half-to-Half

Prediction

Zhizhong Han1,3, Xiyang Wang1,2, Yu-Shen Liu1,2∗, Matthias Zwicker31School of Software, Tsinghua University, Beijing, China

2Beijing National Research Center for Information Science and Technology (BNRist)3Department of Computer Science, University of Maryland, College Park, USA

[email protected], [email protected], [email protected], [email protected]

Abstract

Unsupervised feature learning for point clouds has beenvital for large-scale point cloud understanding. Recen-t deep learning based methods depend on learning globalgeometry from self-reconstruction. However, these methodsare still suffering from ineffective learning of local geome-try, which significantly limits the discriminability of learnedfeatures. To resolve this issue, we propose MAP-VAE toenable the learning of global and local geometry by joint-ly leveraging global and local self-supervision. To enableeffective local self-supervision, we introduce multi-angleanalysis for point clouds. In a multi-angle scenario, wefirst split a point cloud into a front half and a back halffrom each angle, and then, train MAP-VAE to learn to pre-dict a back half sequence from the corresponding front halfsequence. MAP-VAE performs this half-to-half predictionusing RNN to simultaneously learn each local geometry andthe spatial relationship among them. In addition, MAP-VAEalso learns global geometry via self-reconstruction, wherewe employ a variational constraint to facilitate novel shapegeneration. The outperforming results in four shape analy-sis tasks show that MAP-VAE can learn more discriminativeglobal or local features than the state-of-the-art methods.

1. Introduction

Point clouds have become a popular 3D representation inmachine vision, autonomous driving, and augmented reali-ty, because they are easy to acquire and manipulate. There-fore, point cloud analysis has emerged as a crucial prob-lem in the area of 3D shape understanding. With the helpof extensive supervised information, recent deep learning

∗Corresponding Author. This work was supported by National KeyR&D Program of China (2018YFB0505400) and NSF (award 1813583).

based feature learning techniques have achieved unprece-dented results in classification, detection and segmenta-tion [27, 31, 22, 35, 23, 39]. However, supervised learningrequires intense manual labeling effort to obtain supervisedinformation. Therefore, unsupervised feature learning isanattractive alternative and a promising research challenge.

Several studies have tried to adress this challenge [1, 21,4, 34, 33, 42, 32]. To learn the structure of a point cloudwithout additional supervision, these generative models aretrained by self-supervision, such as self-reconstruction[1,4, 6, 42, 33, 32] or distribution approximation [1, 21, 34],which is implemented by auto-encoder or generative ad-versarial networks [8] respectively. To capture finer glob-al structure, some methods [34, 33, 42, 32] first learn localstructure information in point cloud patches based on whichthe global point cloud is then reconstructed. Because oflacking effective and semantic local structure supervision,however, error may accumulate in the local structure learn-ing process, which limits the network’s ability in 3D pointcloud understanding.

To resolve this issue, we propose a novel deep learn-ing model for unsupervised point cloud feature learning bysimultaneously employing effective local and global self-supervision. We introduce multi-angle analysis for pointclouds to mine effective local self-supervision, and com-bine it with global self-supervision under a variational con-straint. Hence we call our model Multi-Angle Point CloudVariational Auto-Encoder (MAP-VAE). Specifically, to em-ploy local self-supervision, MAP-VAE first splits a pointcloud into a front half and a back half under each of sev-eral incrementally varying angles. Then, MAP-VAE per-forms half-to-half prediction to infer a sequence of severalback halves from the corresponding sequence of the com-plementary front halves. Half-to-half prediction aims tocapture the geometric and structural information of localregions on the point cloud through varying angles. More-

Page 2: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

over, by leveraging global self-supervision, MAP-VAE con-ducts self-reconstruction in company with each half-to-halfprediction to capture the geometric and structural informa-tion of the whole point cloud. Self-reconstruction is startedfrom a variational feature space, which enables MAP-VAEto generate new shapes by capturing the distribution infor-mation over training point clouds in the feature space. Insummary, our contributions are as follows:

i) We propose MAP-VAE to perform unsupervised fea-ture learning for point clouds. It can jointly leverageeffective local and global self-supervision to learn fine-grained geometry and structure of point clouds.

ii) We introduce multi-angle analysis for point cloudunderstanding, which provides semantic local self-supervision to learn local geometry and structure.

iii) We provide a novel way to consistently split pointclouds into semantic regions according to view angles,which enables the exploration of the fine-grained dis-criminative information of point cloud regions.

2. Related work

Deep learning models have led to significant progress infeature learning for 3D shapes [12, 11, 14, 13, 17, 18, 9,19, 15, 10]. Here, we focus on reviewing studies on pointclouds. For supervised methods, supervised information,such as shape class labels or segmentation labels, are re-quired to train deep learning models in the feature learningprocess. In contrast, unsupervised methods are designedto mine self-supervision information from point clouds fortraining, which eliminates the need for supervised informa-tion that can be tedious to obtain. We briefly review thestate-of-the-art methods in these two categories as follows.Supervised feature learning.As a pioneering work, Point-Net [26] was proposed to directly learn features from pointclouds by deep learning models. However, PointNet is lim-ited in capturing contextual information among points. Toresolve this issue, various techniques were proposed to es-tablish graph in a local region to capture the relationship a-mong points in the region [31, 22, 35, 23, 39]. Furthermore,multi-scale analysis [27] was introduced to extract more se-mantic features from the local region by separating pointsinto scales or bins, and then, aggregating these features byconcatenation [38] or RNN [25]. These methods require su-pervised information in the feature learning process, whichis different from unsupervised approach in MAP-VAE.Unsupervised feature learning. An intuitive approachto mine self-supervised information is to perform self-reconstruction which first encodes a point cloud into a fea-ture and then decodes the feature back to a point cloud.Such global self-supervision is usually implemented by an

GRU GRU GRU GRU

Global

EncoderLocal Encoder

VariationalReparameterization

Glo

bal

Dec

od

erFront half

Back half

A training sample

GRU GRU GRU

Aggregation branch A

Reconstruction branch R

Prediction branch P

Ag

gre

gat

ion

RN

N

Pred

iction

RN

N

+SF

SB

M

M

UA

UP

SF

M一D

SB一

mFv

mBv

Ti

f f Fv

mFv

hi

Figure 1. The framework of MAP-VAE.

autoencoder network [1, 4, 42, 33, 32]. With the help of ad-versarial training, a different kind of global self-supervisionis employed to train the network to generate plausible pointclouds by learning a mapping from a known distributionto the unknown distribution that the point clouds are sam-pled from [1, 21, 34]. For finer global structure infor-mation, some methods take a step further to jointly em-ploy local structure information captured in local region-s [34, 33, 42, 32]. These methods first learn local structureinformation in point cloud patches by clustering [42], con-ditional point distribution [33], graph convolution [34],orfully connected layers [32], based on which the whole pointcloud is then reconstructed. However, because of lackingeffective and semantic local structure supervision, this pro-cess is prone to error accumulation in the local structurelearning process, which limits the network’s ability in pointcloud understanding. To resolve this issue, MAP-VAE in-troduces multi-angle analysis for point clouds, which pro-vides effective and semantic local self-supervision. MAP-VAE can also simultaneously employ local and global self-supervision, which further differentiates it from others.

3. Overview

To jointly leverage local and global self-supervision tolearn features from point clouds, MAP-VAE simultaneous-ly conducts half-to-half prediction and self-reconstructionby three branches, i.e., which we call aggregation branchA, reconstruction branchR, and prediction branchP, as il-lustrated in Fig. 1. Specifically, branchA and branchPtogether perform the half-to-half prediction while branchAand branchR together perform the self-reconstruction.

A training sampleTi provided to MAP-VAE to learn isformed by a front half sequenceSF , a back half sequenceSB, and an original point cloudM . The corresponding ele-

ments in sequencesSF andSB are a front halfmFv (in red)

and its complementary back halfmBv (in green) which are

obtained by splitting the original point cloudM (in blue)from a specific anglev.

The aggregation branchA encodes the geometry of local

Page 3: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

Front half

Back half

Training sample

....

..1

23

1

2

3

Nea

rest

to v

iew

poit

Geo

des

ic m

ap

Spli

ttin

g

Fro

nt/

bac

k

(a) (c)(b)

+

+

M

mBvmF

v +

Front half

Back half

Training sample

SF

SB

SF

SB

T1

TV

M

M

M

V angles

mFv

mBv

Figure 2. (a) Geodesic splittingM into a front halfmF

v (in red)and a back halfmB

v (in green) from thev-th angle. (b)M is fur-ther split from allV angles located aroundM in clockwise order,where subset withW = 3 out ofV angles (indicated by dotted orsolid line) are selected to establish a half-to-half sequence pair. (c)The training sampleTi from each one ofV angles.

point clouds and their spatial relationship by aggregatingall front halves in sequenceSF in order. It first extractsthe low-level featuref of the original point cloudM andthe low-level featurefF

v of each front halfmFv by a global

encoder and a local encoder, respectively. Then, it learns theangle-specific featurehi of M by aggregating all low-levelfeaturesfF

v using an aggregation RNNUA.The reconstruction branchR performs self-

reconstruction by decoding the learned angle-specificfeaturehi into a point cloudM . This reconstruction isconducted by a global decoderD which tries to generateM as similar as possible toM . In addition,R employs avariational constraint to facilitate novel shape generation.

At the same time, the prediction branchP performs half-to-half prediction by decoding the learned angle-specificfeaturehi into a back half sequenceSB which is pairedwith the corresponding front half sequenceS

F . This pre-diction is conducted by a prediction RNNUP which tries topredict the sequenceSB as similar as possible toSB.

4. Multi-angle splitting

Multi-angle splitting. A key idea in MAP-VAE is a novelmulti-angle analysis for point coulds to mine effective localself-supervision. Intuitively, observing a point cloud fromdifferent angles, explicitly presents the correspondencesand relationships among different shape regions, given asthe correspondence between the front and back halves ofthe shape in each view. Our multi-angle analysis providesmultiple regions (front halves) of the point cloud as in-put, from which the corresponding missing regions (backhalves) need to be inferred. This encourages MAP-VAEto learn a detailed shape representation that facilitates highquality classification, segmentation, shape synthesis, andpoint cloud completion.

We achieve this by splitting a point cloud into a front

(a) (b) (c)

Figure 3. The comparison of front halves (red parts in (a) and(c))split by Euclidean distance and geodesic distance (map in (b)).

and a back half from different angles, where the front halfis the half nearer to the viewpoint than the back half. Thisenables MAP-VAE to observe different semantic parts ofa point cloud, and it also preserves the spatial relationshipamong the parts by incrementally varying angles.

For a point cloudM , we splitM from V different an-gles into front halves (in red) and their complementary backhalves (in green), as shown in Fig. 2 (b), where the view-points are located aroundM on a circle. From thev-thviewpoint,M is split into a front halfmF

v and a back halfmB

v in Fig. 2 (a), wheremFv is formed by theN nearest

points (in red) ofM to the viewpoint whilemBv is formed

by theN farthest points (in green) to the viewpoint.Geodesic splitting. A naive way of finding theN nearestpoints to define a front halfmF

v is to sort all points onMby the Euclidean distance between each point and the view-point. However, on some point clouds, this method may notproduce semantic front halves, since the regions in a fronthalf are not continuous, as demonstrated in Fig. 3 (a). It isimportant to encode semantic front halves, since this wouldhelp MAP-VAE to seamlessly understand the entire surfacefrom a viewpoint under a multi-angle scenario.

To resolve this issue, we leverage the geodesic distanceon the point cloud [2] to sort the points. Specifically, wefirst find a nearest pointu to thev-th viewpoint onM byEuclidean distance. Then, we sort the rest of points onM interms of their geodesic distances to the found nearest pointu, as shown by the geodesic map in Fig. 3 (b). Finally,u

and its nearestN − 1 points form the front halfmFv , while

the farthestN points form the back halfmBv , as illustrated

by the red part and green part in Fig. 3 (c), respectively.Half-to-half sequence pairs. To leverage the correspon-dence between front half and back half and their spatial re-lationship under different angles, we establish a half-to-halfsequence pair starting from each one of theV angles.

Along the circle direction of varying angles, as illustrat-ed by the clockwise green arrow in Fig. 2 (b), we select fronthalvesmF

v and their complementary back halvesmBv from

W out of theV angles. The selectedmFv form a front half

sequenceSF while the complementarymBv form a back

half sequenceSB , whereSF = [mFv |v ∈ [1, V ], |v| = W ],

SB = [mB

v |v ∈ [1, V ], |v| = W ] and each element inSF

corresponds to its complementary element inSB. Thus, a

half-to-half sequence pair(SF ,SB) consists ofSF andSB.To comprehensively observe the point cloud, we select

W angles which uniformly cover the whole shape in each

Page 4: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

half-to-half sequence pair(SF ,SB). As demonstrated bythe dotted lines in Fig. 2 (b), we selectW = 3 angles inorder to form the first(SF ,SB), and then, along the greenarrow, we form the last(SF ,SB) by angles indicated bythe solid lines. Each(SF ,SB) forms a training sampleTi

in company with the point cloudM . Finally, we obtain allV training samples{Ti|i ∈ [1, V ]} fromM in Fig. 2 (c).

5. MAP-VAE

Aggregation branch A. For a training sampleTi con-taining a half-to-half sequence pair(SF ,SB) and the pointcloudM , aggregation branchA encodes the global geome-try of M , local geometry of each one ofW m

Fv in S

F , andthe spatial relationship amongmF

v . Aggregation branchAfirst extracts the geometry of each involved point cloud intoa low-level feature by encoder, and then, aggregates all thelow-level features with their spatiality by aggregation RNNUA. Specifically, we extract the low-level featuref of 2Npoints onM by a global encoder, and the low-level featurefFv of N points onmF

v by a local encoder. Both the glob-al and local encoders employ the same architecture as theencoder in PointNet++ [27], the only difference is the inputnumber of points. Subsequently, aggregation RNNUA ag-gregatesf and allfF

v in W + 1 steps, where we employGRU cell with512 hidden state. Finally, we use the hiddenstate as the angle-specific featurehi of M since the firstfront half inSF is observed starting from thei-th angle.Reconstruction branchR. By decoding the learned angle-specific featurehi, reconstruction branchR tries to generatea point cloudM as similar as possible to the original pointcloudM by a global decoderD. D is formed by 3 fully con-nected layers (1024-2048-6114) and 2 convolutional layers(with 256 and 31×1 kernels each), where batch normaliza-tion is used between every two layers. Here, we prefer EarthMovers distance (EMD) [29] to Chamfer Distance (CD) [5]to evaluate the distance between the reconstructedM andthe originalM , since EMD is more faithful than CD to thevisual quality of point clouds[1]. The EMD distance be-tweenM andM is regarded as the cost of reconstructionto optimize, as defined below, whereφ is a bijection from apointx onM to its corresponding pointφ(x) onM ,

CD = minφ:M→M

x∈M

‖x− φ(x)‖2. (1)

In addition, we employ a variational constraint [20] inreconstruction branchR to facilitate novel shape genera-tion. This is implemented by a variational reparameteri-zation process, as shown in Fig. 1. The variational repa-rameterization transforms the angle-specific featurehi intoanother latent vectorz that roughly follows a unit multi-dimensional Gaussian distribution. After training, branchR can generate a novel shape by sampling a latent vectorfrom the unit Gaussian to the global decoderD.

Specifically, the variational reparameterization first em-ploys fully connected layers to respectively estimate themeanµ and varianceσ for the distribution ofhi. Then,a noise vectorε is sampled from a unit multi-dimensionalGaussian distributionN (0,1), asε ∼ N (0,1) andε ∈R1×Z . Finally, we scale the noiseε byσ and further shift itbyµ, such that the latent vectorz = µ+ ε⊙ σ. The vari-ational reparameterization enables reconstruction branchRto push the distributionq(z|hi) to follow the unit multi-dimensional Gaussian distribution by minimizing the KLdivergence between the distributionq(z|hi) andN (0,1).

Thus, the cost of reconstruction branchR is definedbased on Eq. (1) below, whereα is a balance parameter.

CR = CD + α×KL(q(z|hi)||N (0,1)). (2)

Prediction branch P. Similar to reconstruction branchR,prediction branch P decodes the learned angle-specific fea-turehi to predict the back half sequenceSB correspondingto S

F . BranchP tries to predict a back half sequenceSB

as similar as possible toSB by a prediction RNNUP. Ateach ofW steps,UP predicts one back halfmB

v in the sameorder of elements inSB. This enablesUP to learn the half-to-half correspondence and the spatial relationship amongthe halves ofM . To further push MAP-VAE to compre-hensively understand the point cloud,UP predicts the low-level featurefB

v of each one ofW back halfmBv rather than

the spatial point coordinates ofmBv , which is complemen-

tary to reconstruction branchR. The ground truth low-levelfeaturefB

v of mBv is also extracted by the local encoder in

branchA. Thus, the cost of branchP is defined as follows,

CP =1

v∈[1,V ],|v|=W

‖fBv − fB

v ‖22. (3)

Objective function. For a sampleTi, MAP-VAE is trainedby minimizing all the aforementioned costs of each branch,as defined below, whereβ is a balance parameter.

min CR + β × CP. (4)

After training, MAP-VAE represents the point cloudMas a global featureH by aggregating the angle-specific fea-turehi learned from each sampleTi of M using max pool-ing, such thatH = Pooli∈[1,V ]{hi}.

6. Experimental results and analysis

In this section, we first explore how the parameters in-volved in MAP-VAE affect the discriminability of learnedglobal features in shape classification. Then, MAP-VAE isevaluated in shape classification, segmentation, novel shapegeneration, and point cloud completion by comparing withstate-of-the-art methods.

Page 5: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

EMD:110.7α=0.05

EMD:64.2α=0.01 EMD:68.7α=0.005 EMD:76.3α=0.001

Ground truth

(a)

EMD:113.2α=0.1

(b) (c)

(d) (e) (f)

Figure 4. The point clouds are reconstructed in (b)-(f) under dif-ferentα compared in Table. 2.

EMD:87.2 EMD:69.6 EMD:46.5 EMD:97.7

(b) (c) (d) (e)(a)

Ground truth

W=1 W=3 W=6 W=12

Figure 5. The original point clouds in (a) are reconstructedin (b)-(e) under differentW compared in Table. 3.

Training. We pre-train the global and local encoders inMAP-VAE respectively under the dataset involved in ex-periments in a self-reconstruction task, where the decoderin PointNet++ [27] for segmentation is modified to workwith our encoders to produce three dimensional point co-ordinates in the last layer. After each PointNet++-basedautoencoder is trained, the pre-trained global and local en-coders are fixed for more efficient training of MAP-VAE.

In all experiments, we choose a more challenging wayto train MAP-VAE by all point clouds in multiple shapeclasses of a benchmark rather than a single shape class,where each point cloud has2048 points and each half hasN = 1024 points. In shape classification experiments, wetrain a linear SVM to evaluate the raw discriminability ofthe learned global featureH .

Initially, we employV = 12 angles to analyze a pointcloud and form a training sample byW = 6 angles uni-formly covering the point cloud. We set balance parameterα = 0.01 andβ = 1000 to make each cost in the sameorder of magnitude. We use aZ = 128 dimensional unitGaussian for the variational constraint.Parameter setting. All experiments on parameter effectexploration are conducted under ModelNet10 [37].

We first evaluate howβ affects MAP-VAE by com-paring the results of differentβ candidates including{10, 100, 1000, 10000}. As shown in Table 1, the resultsget better with increasingβ until β = 1000 and degeneratewhenβ is too big. This observation demonstrates a properrange ofβ. We useβ = 1000 in the following experiments.

Then, we evaluate howα affects MAP-VAE by com-paring the results of differentα candidates including{0.1, 0.05, 0.01, 0.005, 0.001}. As shown in Table 2, the

Table 1. The effect ofβ, α = 0.01, W = 6, Z = 128.β 10 100 1000 10000

ACC% 93.72 93.94 94.82 93.72

Table 2. The effect ofα, β = 1000, W = 6, Z = 128.α 0.1 0.05 0.01 0.005 0.001

ACC% 92.62 92.84 94.82 93.39 93.17

Table 3. The effect ofW , α = 0.01, β = 1000, Z = 128.W 1 3 6 12 S-6 R-6

ACC% 92.95 93.39 94.82 93.17 93.39 92.95

Table 4. The effect ofZ, α = 0.01, β = 1000, W = 6.Z 32 64 128 256

ACC% 93.28 94.16 94.82 93.94

results get better with decreasingα until α = 0.01 anddegenerate whenα is too small. This observation demon-strates how enforcing a unit Gaussian distribution on thelatent vectorz too loosely or strictly affects the discrim-inability of learned global features. We also visualize thepoint clouds reconstructed by branchR under differentα,as demonstrated in Fig. 4. We findα affects the recon-structed point clouds in a similar way to how it affects thediscriminability of learned global features. In the followingexperiments, we setα to 0.01.

Subsequently, we explore how the number of anglesW

of in a training sample affects the performance of MAP-VAE, as shown in Table. 3, where several candidateW

including {1, 3, 6, 12} are employed. We findW = 6achieves the best result, where fewer angles provide lesslocal information while more angles increase redundancy.We also observe a similar phenomenon in the reconstruct-ed point clouds shown in Fig.5. In addition, we also ex-plore other ways of distributing theW = 6 angles, suchas continuously (“S-6”) or randomly (“R-6”), respectively.We find our employed uniform placement is the best, sinceeach training sample could cover the whole point cloud. Inthe following experiments, we useW = 6.

Finally, we explore the effect of theZ-dimensional unitGaussian distribution. In Table. 4, we compare the result-s obtained with differentZ, including {32, 64, 128, 256}.The results get better with increasingZ untilZ = 128whiledegenerating whenZ is too big. We believeZ depends onthe number of training samples, and both64 and128 aregood forZ under ModelNet10.Z is set to 128 below.Ablation study. We further explore how each module inMAP-VAE contributes to the performance. As shown inTable. 5, we remove a loss each time to highlight the corre-sponding module. The degenerated results indicate that allelements contribute to the discriminability of learned fea-tures, and self-reconstruction (“NoR”) contributes morethan the half-to-half prediction (“NoP”).

Page 6: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

Body

Wing

Tail

Engine

(a) (b)

Back

Seat

Leg

Arm

Figure 6. We show segmentation results from the airplane class in (a) and the chair class in (b).

Table 5. Ablation study,α = 0.01,β = 1000,W = 6,Z = 128.No R No P No KL All AE VAE Eucli

% 91.63 92.40 93.17 94.82 92.29 93.28 93.61

In addition, we highlight our half-to-half predictionby showing the results obtained only by the pre-trainedglobal encoder in Fig. 1 and this global encoder with avariational constraint (using the same balance weights asMAP-VAE), as shown by “AE” and “VAE”. These resultsshow that half-to-half prediction can help MAP-VAE under-stand point clouds better by leveraging effective local self-supervision. Moreover, the lower result of “Euclid” also in-dicates geodesic distance is superior to Euclidean distanceto obtain semantic front half in the splitting of a point cloud.Shape classification.We evaluate MAP-VAE in shape clas-sification by comparing it with state-of-the-art methods un-der ModelNet40 [37] and ModelNet10 [37]. All the com-pared methods perform unsupervised 3D feature learningwhile using various 3D representations, including voxels,views and point clouds. As shown in Table 6, MAP-VAEobtains the best performance among these methods underModelNet10. We employ the same parameters involvedin Table 5 to produce our result under ModelNet40, whereMAP-VAE also outperforms all point cloud based methods.Although view-based VIPGAN is a little better than ours,it cannot generate 3D shapes. These results indicate thatMAP-VAE learns more discriminative global features forpoint clouds with the ability of leveraging more effectivelocal self-supervision. Note that the results of LGAN, FNetand NSampler are trained under a version of ShapeNet55that contains more than 57,000 3D shapes. However, thereare only 51,679 3D shapes from ShapeNet55 that are avail-able for public download. Therefore, MAP-VAE cannot betrained under the same number of shapes. To perform faircomparison, we use the codes of LGAN and FNet to pro-duce their results under the same shapes in ModelNet, asshown by “LGAN(MN)” and “FNet(MN)”.

Table 6. The classification accuracy (%) comparison among un-supervised 3D feature learning methods under ModelNet40 andModelNet10.α = 0.01,β = 1000,W = 6,Z = 128.

Methods Modality MN40% MN10%T-L Network[7] Voxel 74.40 -Vconv-DAE[30] Voxel 75.50 80.50

3DGAN[36] Voxel 83.30 91.00VSL[24] Voxel 84.50 91.00

VIPGAN[16] View 91.98 94.05LGAN[1] Points 85.70 95.30

LGAN[1](MN) Points 87.27 92.18NSampler[28] Points 88.70 95.30

FNet[40] Points 88.40 94.40FNet[40](MN) Points 84.36 91.85

MRTNet[6] Points 86.40 -3DCapsule[42] Points 88.90 -PointGrow[33] Points 85.80 -PCGAN[21] Points 87.80 -

Our Points 90.15 94.82

Shape segmentation. We evaluate the local featureslearned by MAP-VAE for each point in shape segmentation.The ShapeNet part dataset [26] is employed in this experi-ment, where point clouds in 16 shape classes are involvedto train MAP-VAE with the same parameters in Table 6.

We first extracted the learned feature of each point fromthe second-last layer in the global decoderD. Extractingthe feature of each single point in the decoding procedurerepresents the ability of MAP-VAE to understand shapeslocally at each point. Second, we map the ground truth labelof each point to the reconstructed point cloud by voting 5nearest labels for each reconstructed point. Third, we traina per-point softmax classifier under the training set, and testthe classifier under the test set.

We use the same approach to obtain the results of theautoencoder in LGAN [1], and this autoencoder with a vari-ational constraint. As the comparison shown in Table 7, our

Page 7: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

(a) (b) (c) (d)

Figure 7. Segmentation comparison between “LGAN” (a), “L-GAN1” (b) in Table 7 and MAP-VAE (c). The ground truth isshown in (d). A color in the same row represents a part class.

results significantly outperform “LGAN” and “LGAN1” interms of both mIoU and classification accuracy. We fur-ther visualize the segmentation comparison on four cases inFig. 7. We find that the captured local geometry informationhelps MAP-VAE not only to reconstruct point clouds better,but also to learn more discriminative features for each pointfor better segmentation. Finally, we show 100 segmentationresults in two challenging shape classes, respectively, i.e.,airplane and chair, in Fig. 6. The consistent segmentationresults also justify the good performance of MAP-VAE.Shape generation.Next we demonstrate how to generatenovel shapes using the trained MAP-VAE. Here, we firstsample a noise vector from the employedZ-dimensionalunit Gaussian distribution, and then, convey the noise to theglobal decoderD in branchR in Fig. 1.

Using MAP-VAE trained under ModelNet10 in Table 6,we generate some novel shapes in each of 10 shape class-es in Fig. 10 (a), where we sample 4 noise vectors aroundthe feature center of each shape class to generate 4 shapeclass specific shapes. The generated point clouds are sharpand with high fidelity, where more local geometry detailsare learned. Moreover, we also observe high quality pointclouds in the same shape class from MAP-VAE trained un-der ShapeNet part dataset in Table 7. We generate 100 air-planes using 100 noise vectors sampled around the featurecenter of the airplane class, as demonstrated in Fig. 10 (b).

We further show the point clouds generated by the inter-polated features between two feature centers of two shapeclasses, where the MAP-VAE trained under ModelNet10 inTable 6 is used. The interpolated point clouds are plausibleand meaningful, and smoothly changed from one center toanother center in Fig. 11 (a) and (b). Similar results can beobserved by interpolations between two shapes in the sameclass in Fig. 11 (c) and (d), where the MAP-VAE trainedunder ShapeNet part dataset in Table 7 is used.

(a) (b) (c)

Figure 8. Compared to the three generated class centers underModelNet10, MAP-VAE (in (c)) learns more local geometry de-tails than “LGAN” (in (a)) and “LGAN1” (in (b)) in Table 7, dueto the effective local self-supervision in half-to-half prediction.

(b) (e)(d)(c)(a)

EMD:0.036152 EMD:0.031853 EMD:0.030103 Ground truth

EMD:0.047563 EMD:0.040906 EMD:0.030873 Ground truth

EMD:0.047256 EMD:0.057198 EMD:0.041904 Ground truth

EMD:0.050879 EMD:0.077586 EMD:0.043032 Ground truth

Figure 9. Visual comparison with “LGAN” (in (b)) in Table 7 and“PCN-EMD” (in (c)) in Table 8 for the completion of partial pointclouds (a). MAP-VAE (in (d)) completes more geometry details.

Finally, we visually highlight the advantage of MAP-VAE by the point cloud generated at the feature center of ashape class. As demonstrated in Fig. 8, we compare MAP-VAE with the autoencoder in LGAN[1], and this autoen-coder with a variational constraint. Using the trained de-coder of each method, we generate a point cloud from thefeature center at each of the three shape classes. Comparedto the three class centers in Fig. 8 (a) and Fig. 8 (b), MAP-VAE in Fig. 8 (c) can generate point clouds with more localgeometry details, such as sharp edges of parts.

Page 8: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

Table 7. The segmentation comparison among unsupervised 3Dfeature learning methods under ShapeNet part dataset. The metric ismIoU(%) and per-point classification accuracy(%) on points.α = 0.01,β = 1000,W = 6,Z = 128.

Methods Mean Aero Bag Cap Car Chair Ear Guitar Knife Lamp Laptop Motor Mug Pistol Rocket Skate Table

mIo

U LGAN [1] 57.04 54.13 48.67 62.55 43.24 68.37 58.34 74.27 68.38 53.35 82.62 18.60 75.08 54.70 37.17 46.71 66.39LGAN1 [1] 56.28 52.16 57.85 62.66 42.01 67.66 52.25 75.37 68.63 49.07 81.52 19.20 75.43 54.34 35.09 41.48 65.73

Ours 67.95 62.73 67.08 72.95 58.45 77.09 67.34 84.83 77.07 60.89 90.84 35.82 87.73 64.24 44.97 60.36 74.75

AC

C LGAN [1] 78.24 74.93 84.36 77.02 71.10 78.23 78.34 84.41 78.29 69.05 86.86 67.93 90.42 81.95 68.44 82.27 78.25LGAN1 [1] 77.35 73.64 84.05 75.93 69.82 77.35 77.45 83.72 78.10 68.45 85.85 66.06 89.69 81.43 67.59 81.10 77.33

Ours 87.45 83.50 93.79 86.12 83.28 87.03 88.08 93.15 86.66 79.31 94.89 77.37 98.86 90.51 77.14 93.21 86.25

(a) (b)

Figure 10. High fidelity novel shape generation by MAP-VAE trained under ModelNet10 in (a) and ShapeNet part dataset in (b).

(a)

(b)

(c)

(d)

Figure 11. We show shape interpolation results between two different shape classes under ModelNet10 in (a) and (b), and the shapeinterpolation results between two shapes in the same class under the ShapeNet part dataset in (c) and (d).

Table 8. The completion comparison under airplane and chairclasses in terms of EMD/point,α = 0,β = 1000,W = 12,Z = 0.Class EPN[3] Folding[40] PCN-CD[41] PCN-EMD[41] LGAN[1] Our

Airplane 0.061960 0.156438 0.046637 0.038752 0.0332180.032328Chair 0.076802 0.297427 0.086787 0.068074 0.0559080.055696

Point cloud completion. MAP-VAE can also be used inpoint cloud completion, where we setW = 12 and removethe KL loss for fair comparison. We evaluate our perfor-mance under the training and test sets of partial point cloud-s in two challenging shape classes in [3], i.e., airplane andchair, where we employ the complete point clouds in [26]as ground truth. Since each partial point cloud has differen-t number of points, we resample 2048 points to obtain thefront and back halves. We compare with the state-of-the-artmethods in Table 8. The lowest EMD distance shows thatMAP-VAE outperforms all competitors. In addition, we al-so visually compare the completed point clouds in Fig. 9,

where MAP-VAE completes more local geometry details.

7. Conclusions

We propose MAP-VAE for unsupervised 3D point cloudfeature learning by jointly leveraging local and global self-supervision. MAP-VAE effectively learns local geometryand structure on point clouds from semantic local self-supervision provided by our novel multi-angle analysis.The outperforming results in various applications show thatMAP-VAE successfully learns more discriminative globalor local features for point clouds than state-of-the-art.

Page 9: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

References

[1] P. Achlioptas, O. Diamanti, I. Mitliagkas, and L. J. Guibas.Learning representations and generative models for 3D pointclouds. InThe International Conference on Machine Learn-ing, pages 40–49, 2018.

[2] K. Crane, C. Weischedel, and M. Wardetzky. The heatmethod for distance computation.Communications of theACM, 60(11):90–99, 2017.

[3] A. Dai, C. R. Qi, and M. Nießner. Shape completion us-ing 3D-encoder-predictor cnns and shape synthesis. InIEEEConference on Computer Vision and Pattern Recognition,2017.

[4] H. Deng, T. Birdal, and S. Ilic. PPF-FoldNet: Unsupervisedlearning of rotation invariant 3D local descriptors. InPro-ceedings of European Conference on Computer Vision, vol-ume 11209, pages 620–638, 2018.

[5] H. Fan, H. Su, and L. J. Guibas. A point set generation net-work for 3D object reconstruction from a single image. InIEEE Conference on Computer Vision and Pattern Recogni-tion, pages 2463–2471, 2017.

[6] M. Gadelha, R. Wang, and S. Maji. Multiresolution tree net-works for 3D point cloud processing. InECCV, 2018.

[7] R. Girdhar, D. F. Fouhey, M. Rodriguez, and A. Gupta.Learning a predictable and generative vector representationfor objects. InProceedings of European Conference on Com-puter Vision, pages 484–499, 2016.

[8] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu,D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Gen-erative adversarial nets. InAdvances in Neural InformationProcessing Systems 27, pages 2672–2680. 2014.

[9] Z. Han, X. Liu, Y.-S. Liu, and M. Zwicker. Parts4Feature:Learning 3D global features from generally semantic parts inmultiple views. InIJCAI, 2019.

[10] Z. Han, Z. Liu, J. Han, C. Vong, S. Bu, and C. Chen. Unsu-pervised learning of 3D local features from raw voxels basedon a novel permutation voxelization strategy.IEEE Transac-tions on Cybernetics, 49(2):481–494, 2019.

[11] Z. Han, Z. Liu, J. Han, C.-M. Vong, S. Bu, and C. Chen.Mesh convolutional restricted boltzmann machines for un-supervised learning of features with structure preservationon 3D meshes.IEEE Transactions on Neural Network andLearning Systems, 28(10):2268 – 2281, 2017.

[12] Z. Han, Z. Liu, J. Han, C.-M. Vong, S. Bu, and X. Li. Un-supervised 3D local feature learning by circle convolutionalrestricted boltzmann machine.IEEE Transactions on ImageProcessing, 25(11):5331–5344, 2016.

[13] Z. Han, Z. Liu, C. Vong, Y.-S. Liu, S. Bu, J. Han,and C. Chen. Deep Spatiality: Unsupervised learning ofspatially-enhanced global and local 3D features by deep neu-ral network with coupled softmax.IEEE Transactions onImage Processing, 27(6):3049–3063, 2018.

[14] Z. Han, Z. Liu, C.-M. Vong, Y.-S. Liu, S. Bu, J. Han, andC. Chen. BoSCC: Bag of spatial context correlations for spa-tially enhanced 3D shape representation.IEEE Transactionson Image Processing, 26(8):3707–3720, 2017.

[15] Z. Han, H. Lu, Z. Liu, C.-M. Vong, Y.-S. Liu, M. Zwicker,J. Han, and C. P. Chen. 3D2SeqViews: Aggregating sequen-

tial views for 3d global feature learning by cnn with hierar-chical attention aggregation.IEEE Transactions on ImageProcessing, 28(8):3986–3999, 2019.

[16] Z. Han, M. Shang, Y.-S. Liu, and M. Zwicker. View Inter-Prediction GAN: Unsupervised representation learning for3D shapes by learning global shape memories to support lo-cal view predictions. InAAAI, 2019.

[17] Z. Han, M. Shang, Z. Liu, C.-M. Vong, Y.-S. Liu, M. Zwick-er, J. Han, and C. P. Chen. SeqViews2SeqLabels: Learning3D global features via aggregating sequential views by rn-n with attention. IEEE Transactions on Image Processing,28(2):1941–0042, 2019.

[18] Z. Han, M. Shang, X. Wang, Y.-S. Liu, and M. Zwick-er. Y2Seq2Seq: Cross-modal representation learning for 3Dshape and text by joint reconstruction and prediction of viewand word sequences. InAAAI, 2019.

[19] Z. Han, X. Wang, C.-M. Vong, Y.-S. Liu, M. Zwicker, andC. P. Chen. 3DViewGraph: Learning global features for 3dshapes from a graph of unordered views with attention. InIJCAI, 2019.

[20] D. P. Kingma and M. Welling. Auto-encoding variationalbayes.CoRR, abs/1312.6114, 2013.

[21] C.-L. Li, M. Zaheer, Y. Zhang, B. Poczos, and R. Salakhut-dinov. Point cloud GAN.CoRR, abs/1810.05795, 2018.

[22] J. Li, B. M. Chen, and G. H. Lee. SO-Net: Self-organizingnetwork for point cloud analysis. InProceedings of the IEEEConference on Computer Vision and Pattern Recognition,pages 9397–9406. IEEE Computer Society, 2018.

[23] Y. Li, R. Bu, M. Sun, W. Wu, X. Di, and B. Chen. PointCN-N: Convolution on x-transformed points. InNeurIPS, pages828–838, 2018.

[24] S. Liu, C. L. Giles, and A. G. O. II. Learning a hierarchicallatent-variable model of 3D shapes. In2018 InternationalConference on 3D Vision (3DV), 2018.

[25] X. Liu, Z. Han, Y. Liu, and M. Zwicker. Point2Sequence:Learning the shape representation of 3D point clouds with anattention-based sequence to sequence network.AAAI, 2019.

[26] C. R. Qi, H. Su, K. Mo, and L. J. Guibas. PointNet: Deeplearning on point sets for 3D classification and segmentation.In IEEE Conference on Computer Vision and Pattern Recog-nition, 2017.

[27] C. R. Qi, L. Yi, H. Su, and L. J. Guibas. PointNet++:Deep hierarchical feature learning on point sets in a metricspace. InAdvances in Neural Information Processing Sys-tems, pages 5105–5114, 2017.

[28] E. Remelli, P. Baque, and P. Fua. NeuralSampler: Eu-clidean point cloud auto-encoder and sampler.CoRR, ab-s/1901.09394, 2019.

[29] Y. Rubner, C. Tomasi, and L. J. Guibas. The earth mover’sdistance as a metric for image retrieval.International Jour-nal of Computer Vision, 40(2):99–121, 2000.

[30] A. Sharma, O. Grau, and M. Fritz. VConv-DAE: Deep volu-metric shape learning without object labels. InProceedingsof European Conference on Computer Vision, pages 236–250, 2016.

[31] Y. Shen, C. Feng, Y. Yang, and D. Tian. Mining point cloudlocal structures by kernel correlation and graph pooling. In

Page 10: Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning ...zwicker/publications/MultiAngle...Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds from

Proceedings of the IEEE Conference on Computer Visionand Pattern Recognition, pages 4548–4557, 2018.

[32] M. Shoef, S. Fogel, and D. Cohen-Or. PointWise:an un-supervised point-wise feature learning network.CoRR, ab-s/1901.04544, 2019.

[33] Y. Sun, Y. Wang, Z. Liu, J. E. Siegel, and S. E. Sarma. Point-Grow: Autoregressively learned point cloud generation withself-attention.CoRR, abs/1810.05591, 2018.

[34] D. Valsesia, G. Fracastoro, and E. Magli. Learning localizedgenerative models for 3D point clouds via graph convolu-tion. In International Conference on Learning Representa-tions, 2019.

[35] Y. Wang, Y. Sun, Z. Liu, S. E. Sarma, M. M. Bronstein, andJ. M. Solomon. Dynamic graph CNN for learning on pointclouds.CoRR, abs/1801.07829, 2018.

[36] J. Wu, C. Zhang, T. Xue, B. Freeman, and J. Tenenbaum.Learning a probabilistic latent space of object shapes via 3Dgenerative-adversarial modeling. InAdvances in Neural In-formation Processing Systems, pages 82–90. 2016.

[37] Z. Wu, S. Song, A. Khosla, F. Yu, L. Zhang, X. Tang, andJ. Xiao. 3D ShapeNets: A deep representation for volumetricshapes. InProceedings of IEEE Conference on ComputerVision and Pattern Recognition, pages 1912–1920, 2015.

[38] S. Xie, S. Liu, Z. Chen, and Z. Tu. Attentional shapecon-textnet for point cloud recognition. InProceedings of theIEEE Conference on Computer Vision and Pattern Recogni-tion, pages 4606–4615, 2018.

[39] Y. Xu, T. Fan, M. Xu, L. Zeng, and Y. Qiao. SpiderCNN:Deep learning on point sets with parameterized convolution-al filters. InECCV, volume 11212, pages 90–105, 2018.

[40] Y. Yang, C. Feng, Y. Shen, and D. Tian. FoldingNet: Pointcloud auto-encoder via deep grid deformation. InIEEE Con-ference on Computer Vision and Pattern Recognition, 2018.

[41] W. Yuan, T. Khot, D. Held, C. Mertz, and M. Hebert. PCN:Point completion network. InProceedings of 2018 Interna-tional Conference on 3D Vision, 2018.

[42] Y. Zhao, T. Birdal, H. Deng, and F. Tombari. 3D point-capsule networks.CoRR, abs/1812.10775, 2018.