13
W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 1 Improving Heterogeneous Face Recognition with Conditional Adversarial Networks Wuming Zhang 1 [email protected] Zhixin Shu 2 [email protected] Dimitris Samaras 2 [email protected] Liming Chen 1 [email protected] 1 Laboratory LIRIS Ecole Centrale de Lyon Ecully, France 2 Computer Vision Lab Stony Brook University Stony Brook, NY, USA Abstract Heterogeneous face recognition between color image and depth image is a much desired capacity for real world applications where shape information is looked upon as merely involved in gallery. In this paper, we propose a cross-modal deep learning method as an effective and efficient workaround for this challenge. Specifically, we be- gin with learning two convolutional neural networks (CNNs) to extract 2D and 2.5D face features individually. Once trained, they can serve as pre-trained models for another two- way CNN which explores the correlated part between color and depth for heterogeneous matching. Compared with most conventional cross-modal approaches, our method ad- ditionally conducts accurate depth image reconstruction from single color image with Conditional Generative Adversarial Nets (cGAN), and further enhances the recognition performance by fusing multi-modal matching results. Through both qualitative and quan- titative experiments on benchmark FRGC 2D/3D face database, we demonstrate that the proposed pipeline outperforms state-of-the-art performance on heterogeneous face recognition and ensures a drastically efficient on-line stage. 1 Introduction For decades, face recognition (FR) from color images has achieved substantial progress and forms part of an ever-growing number of real world applications, such as video surveil- lance, people tagging and virtual/augmented reality systems [3, 32, 40]. With the increasing demand for recognition accuracy under unconstrained conditions, the weak points of 2D based FR methods become apparent: as an imaging-based representation, color image is quite sensitive to numerous external factors, such as lighting variations and makeup patterns. Therefore, 3D based FR techniques [5, 7, 8] have recently emerged as a remedy because they take into consideration the intrinsic shape information of faces which is more robust while dealing with these nuisance factors. Moreover, the complementary strengths of color and depth data allow them to jointly work and gain further improvement. c 2017. The copyright of this document resides with its authors. It may be distributed unchanged freely in print or electronic forms. arXiv:1709.02848v2 [cs.CV] 13 Sep 2017

Improving Heterogeneous Face Recognition with Conditional

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 1

Improving Heterogeneous Face Recognitionwith Conditional Adversarial Networks

Wuming Zhang1

[email protected]

Zhixin Shu2

[email protected]

Dimitris Samaras2

[email protected]

Liming Chen1

[email protected]

1 Laboratory LIRISEcole Centrale de LyonEcully, France

2 Computer Vision LabStony Brook UniversityStony Brook, NY, USA

Abstract

Heterogeneous face recognition between color image and depth image is a muchdesired capacity for real world applications where shape information is looked uponas merely involved in gallery. In this paper, we propose a cross-modal deep learningmethod as an effective and efficient workaround for this challenge. Specifically, we be-gin with learning two convolutional neural networks (CNNs) to extract 2D and 2.5D facefeatures individually. Once trained, they can serve as pre-trained models for another two-way CNN which explores the correlated part between color and depth for heterogeneousmatching. Compared with most conventional cross-modal approaches, our method ad-ditionally conducts accurate depth image reconstruction from single color image withConditional Generative Adversarial Nets (cGAN), and further enhances the recognitionperformance by fusing multi-modal matching results. Through both qualitative and quan-titative experiments on benchmark FRGC 2D/3D face database, we demonstrate thatthe proposed pipeline outperforms state-of-the-art performance on heterogeneous facerecognition and ensures a drastically efficient on-line stage.

1 IntroductionFor decades, face recognition (FR) from color images has achieved substantial progress andforms part of an ever-growing number of real world applications, such as video surveil-lance, people tagging and virtual/augmented reality systems [3, 32, 40]. With the increasingdemand for recognition accuracy under unconstrained conditions, the weak points of 2Dbased FR methods become apparent: as an imaging-based representation, color image isquite sensitive to numerous external factors, such as lighting variations and makeup patterns.Therefore, 3D based FR techniques [5, 7, 8] have recently emerged as a remedy because theytake into consideration the intrinsic shape information of faces which is more robust whiledealing with these nuisance factors. Moreover, the complementary strengths of color anddepth data allow them to jointly work and gain further improvement.

c© 2017. The copyright of this document resides with its authors.It may be distributed unchanged freely in print or electronic forms.

arX

iv:1

709.

0284

8v2

[cs

.CV

] 1

3 Se

p 20

17

Citation
Citation
{Azeem, Sharif, Raza, and Murtaza} 2014
Citation
Citation
{Tan, Chen, Zhou, and Zhang} 2006
Citation
Citation
{Zhao, Chellappa, Phillips, and Rosenfeld} 2003
Citation
Citation
{Bowyer, Chang, and Flynn} 2006
Citation
Citation
{Corneanu, Sim{ó}n, Cohn, and Guerrero} 2016
Citation
Citation
{Ding and Tao} 2016
Page 2: Improving Heterogeneous Face Recognition with Conditional

2 W. ZHANG ET AL.: IMPROVING HFR WITH CGAN

Figure 1: Overview of the proposed CNN models for heterogeneous face recognition. Notethat (1) depth recovery is conducted only for testing; (2) the final joint recognition may ormay not include color based matching, depending on the specific experiment protocol.

However, depth data is not always accessible in real-life conditions due to its special re-quirements for optical instruments and acquisition environment. Likewise, other challengesremain as well, including the real-time registration and preprocessing for depth images. Animportant question then naturally arises: can we design a recognition pipeline where depthimages are only registered in gallery while still providing significant information for the iden-tification of unseen color images? To cope with this problem, heterogeneous face recognition(HFR) [13, 33, 41] has been proposed as a reasonable workaround. As a worthwhile trade-offbetween purely 2D and 3D based method, HFR adopts both color and depth data for train-ing and gallery set while the online probe set will simply contains color images. Under thismechanism, a HFR framework can take full advantage of both color and depth informationat the training stage to reveal the correlation between them. Once learned, this cross-modalcorrelation makes it possible to conduct heterogeneous matching between preloaded depthimages in gallery and color images digitally captured in real time.

Beyond the above-mentioned mechanism, in this paper we take a further look at ourconstraint on the use of depth image. Note that all difficulties, which hinder us from availingourselves of depth information in probe set, come from the acquisition and registration of3D data. Intuitively, these problems can be immediately solved if we can reconstruct depthimage from color image accurately and efficiently. Despite many existing work on shaperecovery from single image, most of them rely on 3D model fitting which is time-consumingand can be prone to lack accuracy when landmarks are not precisely located. Thanks to theextremely rapid development of generative models, especially the Generative AdversarialNetwork (GAN) [10] and its conditional variation (cGAN) [22] which are introduced quiterecently, we implement an end-to-end depth face recovery with cGAN to enforce the realisticimage generation. Furthermore, the recovered depth information enables a straightforwardcomparison in 2.5D space.

A flowchart of the proposed method is illustrated in Fig. 1, and we list our contributionsas follows:

• A novel depth face recovery method based on cGAN and Auto-encoder with skip

Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Toderici, Passalis, Zafeiriou, Tzimiropoulos, Petrou, Theoharis, and Kakadiaris} 2010
Citation
Citation
{Zhao, Zhang, Evangelopoulos, Huang, Shah, Wang, Kakadiaris, and Chen} 2013
Citation
Citation
{Goodfellow, Pouget-Abadie, Mirza, Xu, Warde-Farley, Ozair, Courville, and Bengio} 2014
Citation
Citation
{Mirza and Osindero} 2014
Page 3: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 3

connections which greatly improves the quality of reconstructed depth images.

• We first train two discriminative CNNs individually for a two-fold purpose: to extractfeatures of color image and depth image, and to provide pre-trained models for thecross-modal 2D/2.5D CNN model.

• A novel heterogeneous face recognition pipeline which fuses multi-modal matchingscores to achieve state-of-the-art performance.

2 Related Work

2.1 3D Face Reconstruction3D face reconstruction from single/multiple images or stereo video has been a challengingtask due to its nonlinearity and ill-posedness. A number of prevailing approaches addressedthis problem based on shape-subspace projections, where a set of 3D prototypes are fittedby adjusting corresponding parameters to a given 2D image and most of them were derivedfrom 3DMM [4] and Active Appearance Models [21]. Alternative models were afterwardsproposed as well which follow the similar processing pipeline by fitting 3D models to 2Dimages through various face collections or prior knowledge. For example, Gu and Kanade[11] fit surface 3D points and related textures together with the pose and deformation esti-mation. Kemelmacher-Shlizerman et al. [17] considered the input image as a guide with asingle reference model to achieve 3D reconstruction. In recent work of Liu et al. [20], twosets of cascaded regressors are implemented and correlated via a 3D-2D mapping iterativelyto solve face alignment and 3D face reconstruction simultaneously. Likewise, using genericmodel remains a decent solution as well for 3D face reconstruction from stereo videos, aspresented in [6, 9, 23]. Despite of strikingly accurate reconstruction result reported in theabove researches, the drawback of relying on single or a large number of well-aligned 3Dtraining data is observed and even enlarged here, because as far as we know 3D prototypesare necessary for almost all reconstruction approaches.

2.2 2D-3D Heterogeneous Face RecognitionAs a pioneer and cornerstone for numerous subsequent 3D Morphable Model (3DMM) basedmethods, Blanz and Vetter [4] built this statistical model by merging a branch of 3D facemodels and then densely fit it to a given facial image for further matching. Toderici et al.[33] located some predefined key landmarks on the facial images in different poses, and thenroughly align them to a frontal 3D model to achieve recognition target; Riccio and Duge-lay [26] also established a dense correspondence between the 2D probe and the 3D galleryusing geometric invariants across face region. Following this framework, a pose-invariantasymmetric 2D-3D FR approach [39] was proposed which conducts a 2D-2D matching bysynthesizing 2D image from corresponding 3D models towards the same pose as a givenprobe sample. This approach was further extended and compared with work of Zhao et al.[41] as a benchmarking asymmetric 2D-3D FR system, a complete version of their workwas recently released in [16]. Though the above models achieved satisfactory performance,unfortunately they all suffer from high computational cost and long convergence process ow-ing to considerable complexity of pose synthesis, and their common assumption that accuratelandmark localization in facial images was fulfilled turns out to be another tough topic. More

Citation
Citation
{Blanz and Vetter} 2003
Citation
Citation
{Matthews, Xiao, and Baker} 2007
Citation
Citation
{Gu and Kanade} 2006
Citation
Citation
{Kemelmacher-Shlizerman and Basri} 2011
Citation
Citation
{Liu, Zeng, Zhao, and Liu} 2016
Citation
Citation
{Chowdhury, Chellappa, Krishnamurthy, and Vo} 2002
Citation
Citation
{Fidaleo and Medioni} 2007
Citation
Citation
{Park and Jain} 2007
Citation
Citation
{Blanz and Vetter} 2003
Citation
Citation
{Toderici, Passalis, Zafeiriou, Tzimiropoulos, Petrou, Theoharis, and Kakadiaris} 2010
Citation
Citation
{Riccio and Dugelay} 2007
Citation
Citation
{Zhang, Huang, Wang, and Chen} 2012
Citation
Citation
{Zhao, Zhang, Evangelopoulos, Huang, Shah, Wang, Kakadiaris, and Chen} 2013
Citation
Citation
{Kakadiaris, Toderici, Evangelopoulos, Passalis, Chu, Zhao, Shah, and Theoharis} 2016
Page 4: Improving Heterogeneous Face Recognition with Conditional

4 W. ZHANG ET AL.: IMPROVING HFR WITH CGAN

recently, learning based approaches have significantly increased on 2D/3D FR. Huang et al.[13] projected the proposed illuminant-robust feature OGM onto the CCA space to maxi-mize the correlation between 2D/3D features; instead, Wang et al. [35] combined RestrictedBoltzmann Machines (RBMs) and CCA/kCCA to achieve this goal. The work of Jin et al.[15], called MSDA based on Extreme Learning Machine (ELM) as aforementioned, aimsat finding a common discriminative feature space revealing the underlying relationship be-tween different views. These approaches take well advantage of learning model, but wouldencounter weakness when dealing with non-linear manifold representations.

3 Depth Face ReconstructionThe target of taking a random color face image to recover its counterpart in depth space isrealized in this section. We first formulate our problem by adapting it to the background ofcGAN, then the detailed architecture design is described and discussed.

3.1 Problem FormulationFirst proposed in [10], GAN has achieved impressive results in a wide variety of generativetasks. The core idea of GAN is to train two neural networks, which respectively representthe generator G and the discriminator D, to proceed a game-theoretic tussle between oneanother. Given the samples x from the real data distribution pdata(x) and random noise zsampled from a noise distribution pz(z), the discriminator aims to distinguish between realsamples x and fake samples which are mapped from z by the generator, while the generatoris tasked with maximally confusing the discriminator. The objective can thus be written as:

LGAN(G,D) = Ex∼pdata(x)[logD(x)]+Ez∼pz(z)[log(1−D(G(z)))] (1)

where E denotes the empirical estimate of expected value of the probability. To optimizethis loss function, we aim to minimize its value for G and maximize it for D in an adversarialway, i.e. minG maxDLGAN(G,D).

The advantage of GAN is that realistic images can be generated from noise vectors withrandom distribution, which is crucially important for unsupervised learning. However, notethat in our face recovery scenario, training data contains image pairs {x,y} where x andy refer to the depth and color faces respectively with a one-to-one correspondence betweenthem. The fact that y can be involved in the model as a prior for generative task leads us to theconditional variant of GAN, namely cGAN [22]. Specifically, we condition the observationsy on both the discriminator and the generator, the objective of cGAN extends (1) to:

LcGAN(G,D) = Ex,y∼pdata(x,y)[logD(x,y)]+Ez∼pz(z),y∼pdata(y)[log(1−D(G(z|y),y))] (2)

Moreover, to ensure the pixel-wise similarity between image generation outputs G(z|y)and ground truth x, we subsequently impose a reconstruction constraint on the generator inthe form of L1 distance between them:

LL1(G) = Ex,y∼pdata(x,y),z∼pz(z)[‖x−G(z|y)‖1] (3)

The comprehensive objective is formulated with a minmax value function on the abovetwo losses where the scalar η is used for balancing them:

minG

maxD

[LcGAN(G,D)+ηLL1(G)] (4)

Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Wang, Ly, Guo, and Kambhamettu} 2014
Citation
Citation
{Jin, Cao, Ruan, and Wang} 2014
Citation
Citation
{Goodfellow, Pouget-Abadie, Mirza, Xu, Warde-Farley, Ozair, Courville, and Bengio} 2014
Citation
Citation
{Mirza and Osindero} 2014
Page 5: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 5

(a) Workflow of cGAN

(b) Generator

(c) DiscriminatorFigure 2: The mechanism and architecture of cGAN. In Fig. 2b, the noise variable z presentsitself under the the form of dropout layers, while the black arrows portray the skip connec-tions. All convolution and deconvolution layers are with filter size 4×4 and 1-padding, n ands represent the number of output channels and stride value, respectively. (Best view in color)

Note that the cGAN itself can hardly generate specified images and using only LL1(G)causes blurring, this joint loss successfully leverages the complementary strengths of them.

3.2 CGAN Architecture

We adapt our cGAN architecture from that in [14] which achieved particularly impressiveresults in image-to-image translation task. A detailed description of this model is illustratedin Fig. 2 and some key features are discussed below.

Generator: As a standard generative model, the architectures of auto-encoder (AE) [12]and its variants [18, 27, 34] are widely adopted as G for past cGANs. However, the draw-back of conventional AEs is obvious: due to their dimensionality reduction capacity, a largeportion of low-level information, such as precise localization, is compressed when an imagepasses through layers in the encoder. To cope with this lossy compression problem, we fol-low the idea of U-Net [28] by adding skip connections which forwards directly the featuresfrom encoder layers to decoder layers that are on the same ’level’, as shown in Fig. 2b.

Discriminator: Consistent with Isola et al. [14], we adopt PatchGAN for the discrim-inator. Within this pattern, no fully connected layers are implemented and D outputs a 2Dimage where each pixel represents the prediction result with respect to the correspondingpatch on original image. All pixels are then averaged to decide whether the input image is’real’ or ’fake’. Compared with pixel-level prediction, PatchGAN efficiently concentrates onlocal patterns while the global low-frequency correctness is enforced by L1 loss in (3).

Optimization: The optimization for cGAN is performed by following the standardmethod [10]: the mini-batch SGD and the Adam solver are applied to optimize G and Dalternately (as depicted by arrows with different colors in Fig. 2a).

Citation
Citation
{Isola, Zhu, Zhou, and Efros} 2016
Citation
Citation
{Hinton and Salakhutdinov} 2006
Citation
Citation
{Kingma and Welling} 2013
Citation
Citation
{Rifai, Vincent, Muller, Glorot, and Bengio} 2011
Citation
Citation
{Vincent, Larochelle, Lajoie, Bengio, and Manzagol} 2010
Citation
Citation
{Ronneberger, Fischer, and Brox} 2015
Citation
Citation
{Isola, Zhu, Zhou, and Efros} 2016
Citation
Citation
{Goodfellow, Pouget-Abadie, Mirza, Xu, Warde-Farley, Ozair, Courville, and Bengio} 2014
Page 6: Improving Heterogeneous Face Recognition with Conditional

6 W. ZHANG ET AL.: IMPROVING HFR WITH CGAN

Figure 3: Training procedure of the cross-modal CNN model. Models in the dashed box arepre-trained using 2D and 2.5D face images individually.

4 Heterogeneous Face Recognition

The reconstruction of depth faces from color images enables us to maximally leverage shapeinformation in both gallery and probe, which means we can individually learn a CNN modelto extract discriminative features for depth images and transform the initial cross-modalproblem into a multi-modal one. However, the heterogeneous matching remains anotherchallenge in our work, below we demonstrate how this problem is formulated and tackled.

Unimodal learning. The last few years witnessed a surge of interest and success inFR with deep learning [24, 30, 31]. Following the basic idea of stacking convolution-convolution-pooling (C-C-P) layers in [19], we train from scratch two CNNs for color andgrayscale images on CASIA-WebFace [37] and further fine-tune the grayscale based modelwith our own depth images. These two models serve two purposes: to extract 2D and 2.5Dfeatures individually, and to offer pre-trained models for the ensuing cross-modal learning.

Cross-modal learning. Once a pair of unimodal models for both views are trained, themodal-specific representations, {X ,Y}, can be obtained after the last fully connected layers.Note that each input for the two-stream cross-modal CNN is a 2D+2.5D image pair withidentity correspondence, it is reasonable to have an intuition that X and Y share commonpatterns which help to classify them as the same class. This connection essentially reflectsthe nature of cross-modal recognition, and was investigated in [13, 35, 36].

In order to explore this shared and discriminative feature, a joint supervision is requiredto enforce both correlation and distinctiveness simultaneously. For this purpose, we applytwo linear mappings following X and Y , denoted by MX and MY . First, to ensure the correla-tion between new features, they are enforced to be as close as possible, which is constrainedby minimizing their distance in feature space:

Lcorr =n

∑i=1‖MX Xi−MYYi‖2

F (5)

where n denotes the size of mini-batch and ‖ · ‖F represents the Frobenius norm.

If we only use the above loss supervision signal, the model will simply learn zero map-pings for MX and MY because the correlation loss will stably be 0 in this case. To avoid thistricky situation, we average the two outputs to obtain a new feature on which the classifica-

Citation
Citation
{Parkhi, Vedaldi, and Zisserman} 2015
Citation
Citation
{Sun, Liang, Wang, and Tang} 2015
Citation
Citation
{Taigman, Yang, Ranzato, and Wolf} 2014
Citation
Citation
{Lee, Chen, Tseng, and Lai} 2016
Citation
Citation
{Yi, Lei, Liao, and Li} 2014
Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Wang, Ly, Guo, and Kambhamettu} 2014
Citation
Citation
{Wang, Lin, Lu, Feng, etprotect unhbox voidb@x penalty @M {}al.} 2016
Page 7: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 7

Databases Training Set Test setBU3D [38] Bosphorus [29] CASIA-3D [1] FRGC Ver2.0 [25]

# Persons 100 105 123 466# Images 2500 2896 1845 4003

Conditions E E EI EI

Table 1: Database overview. E and I are short for expressions and illuminations, respectively.

tion loss is computed. The ultimate objective function is formulated as follows:

Lh f r = Lso f tmax +λLCorr

=−n

∑i=1

logeW T

ci(MX Xi+MY Yi)/2+bci

∑mj=1 eW T

j (MX Xi+MY Yi)/2+b j+λ

n

∑i=1‖MX Xi−MYYi‖2

F

where ci represents the ground truth class label of ith image pair, the scalar λ denotes theweight for correlation loss.

Fusion. To highlight the effectiveness of the proposed method, we adopt the cosinesimilarity of 4096-d hidden layer features as matching scores. As for the score fusion stage,all scores are normalized to [0,1] and fused by a simple sum rule.

5 Experimental ResultsTo intuitively demonstrate the effectiveness of the proposed method, we conduct extensiveexperiments for 2D/2.5D HFR on the benchmark 2D/2.5D face database. Besides the recon-structed 2.5D depth image, our method also outperforms state-of-the-art performance usingonly 2.5D images instead of holistic 3D face models.

5.1 Dataset CollectionCollecting 2D/2.5D image pairs presents itself as a primary challenge when considering deepCNN as a learning pipeline. Unlike the tremendous boost in dataset scale of 2D face images,massive 3D face data acquisition still remains a bottleneck for the development and practicalapplication of 3D based FR techniques, from which our work is partly motivated.

Databases: As listed in Table 1, three large scale and publicly available 3D face databasesare gathered as training set and the performance is evaluated on another dataset, which im-plies that there was no overlap between training and test set and the generalization capacityof the proposed method is evaluated as well. Note that the attribute values only concern thedata used in our experiments, for example, scans with large pose variations in CASIA-3Dare not included here.

Preprocessing: To generate 2.5D range image from original 3D shape, we either proceeda direct projection if the point cloud is pre-arranged in grids (Bosphorus/FRGC) or adopt asimple Z-buffer algorithm (BU3D/CASIA3D). Furthermore, to ensure that all faces are of thesimilar scale, we resize and crop the original image pairs to 128×128 while fixing their inter-ocular distance to a certain value. Especially, to deal with the missing holes and unwantedbody parts (shoulder for example) in raw data of FRGC, we first locate the face based on 68automatically detected landmarks [2], and then apply a linear interpolation to approximatethe default value of each hole pixel by averaging its non-zero neighboring points.

Citation
Citation
{Yin, Wei, Sun, Wang, and Rosato} 2006
Citation
Citation
{Savran, Aly{ü}z, Dibeklio{§}lu, {Ç}eliktutan, G{ö}kberk, Sankur, and Akarun} 2008
Citation
Citation
{CAS} 2004
Citation
Citation
{Phillips, Flynn, Scruggs, Bowyer, Chang, Hoffman, Marques, Min, and Worek} 2005
Citation
Citation
{Asthana, Zafeiriou, Cheng, and Pantic} 2014
Page 8: Improving Heterogeneous Face Recognition with Conditional

8 W. ZHANG ET AL.: IMPROVING HFR WITH CGAN

(a) (b)Figure 4: Qualitative reconstruction results of FRGC samples with varying illuminations andexpressions. Fig. 4a: correctly recovered samples. Fig . 4b: wrongly recovered samples.

5.2 Implementation detailsAll images are normalized before being fed to the network by subtracting from each channelits mean value over all training data. With regards to the choice of hyperparameters, we adoptthe following setting: in cGAN, the learning rate µcGAN is set to 0.0001 and the weight forL1 norm η is 500; in cross-modal CNN model, the learning rate for training from scratch µ

begins with 1 and is divided by 5 every 10 epochs while the learning rate during fine-tuningµ f t is 0.001; for both models, the momentum m is initially set as 0.5 until it is increased to0.9 at the 10th epoch; the weight for correlation loss λ is set to 0.6.

5.3 Reconstruction ResultsThe reconstruction results obtained for color images in FRGC are illustrated in Fig. 4. Sam-ples from different subjects across expression and illumination variations are shown from leftto right. They thereby give hints on the generalization ability of the proposed method. Foreach sample we first portray the original color image with its ground truth depth image, fol-lowed by the reconstructed results whereby we demonstrate the effectiveness and necessityof each constraint in the joint objective. In addition, some samples with low reconstructionquality are depicted in Fig. 4b as well.

Being consistently similar with the ground truth, the reconstruction results with jointloss in Fig. 4a intuitively demonstrate the strength of cGAN. The recovered depth faces holdtheir accuracy and realistic property irrespective of lighting and expression variations in theoriginal RGB images. Furthermore, when we take an observation of the two reconstructionresults in 3rd and 4th rows, the comparison implies that: 1) using only L1 loss will lead toblurry results because the model tends to average all plausible values, especially for regionscontaining high-level information like edges; 2) using only cGAN loss can achieve slightlysharper results, but suffers from noise. These results provide an evidence that the imple-mentation of joint loss is beneficial and important for obtaining a ’true’ and accurate output.Meanwhile, our model encounters some problems while dealing with extreme cases, such as

Page 9: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 9

Protocol Methods Rank-1 Recognition Accuracy2D 2.5D 2D/2.5D Fusion

Jin et al. [15] MSDA+ELM [15] - - 0.9680 -Ours - 0.9573 0.9603 0.9698

Wang et al. [35] GRBM+rKCCA [35] - - 0.9600 -Ours - 0.9529 0.9714 0.9745

Huang et al. [13] OGM [13] 0.9390 - 0.9404 0.9537Ours 0.9755 0.9609 0.9688 0.9792

Table 2: Comparison of recognition accuracy on FRGC under different protocols.

λ 0 0.2 0.4 0.6 0.8 1 1.2Accuracy 0.9245 0.9481 0.9600 0.9688 0.9577 0.8851 0.7333

Table 3: 2D/2.5D HFR accuracy with varying λ under protocol of [13].

thick beard, wide opened mouth and extremely dark shadows as displayed in Fig. 4b. Theerrors are principally due to few training samples with these cases.

5.4 2D-3D Asymmetric FR

We conduct the quantitative experiments on FRGC which has held the field as one of themost commonly used benchmark dataset over the last decade. In contrast with unimodal FRexperiments, very few attempts have been made on 2D/3D asymmetric FR. For convenienceof comparison, three recent and representative protocols reported respectively in [15], [13]and [35] are followed. These protocols mainly differ in gallery and probe setting, includingsplitting and modality setup. For example, the gallery set in [35] solely contains depthimages, when compared with their work, our experiment will subsequently exclude 2D basedmatching to respect this protocol.

The comparison results are shown in Table 2, through which we could gain the obser-vation that the proposed cross-modal CNN outperforms state-of-the-art performance whilefusing 2.5D matching into HFR with reconstructed depth image further helps improve theperformance effectively. Moreover, the proposed method is advantageous in its 3D-free re-construction capacity and efficiency. To the best of our knowledge, this is the first time toinvestigate a 2.5D face recovery approach which is free of any 3D prototype models. De-spite nearly 20 hours for the whole training and fine-tuning procedure, it takes only 1.6 msto complete an online forward pass per image on a single NVIDIA GeForce GTX TITAN XGPU and is therefore capable of satisfying the real-time processing requirement.

Effect of hyperparameter λ . An extended analysis is made to explore the role of soft-max loss and correlation loss. We take the protocol in [13] as a standard and vary the weightfor correlation loss λ each time. As shown in Table 3, the performance will remain largelystable across a range of λc between 0.4 and 0.8. When we set λ = 0 instead of 0.6, whichmeans correlation loss is not involved while training, the network can still learn valuablefeatures with a recognition rate decrease of 4.43%. However, along with the increase of λ ,the performance drops drastically, which implies that a too strong constraint on correlationloss could backfire by causing a negative impact on softmax loss.

Citation
Citation
{Jin, Cao, Ruan, and Wang} 2014
Citation
Citation
{Jin, Cao, Ruan, and Wang} 2014
Citation
Citation
{Wang, Ly, Guo, and Kambhamettu} 2014
Citation
Citation
{Wang, Ly, Guo, and Kambhamettu} 2014
Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Jin, Cao, Ruan, and Wang} 2014
Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Citation
Citation
{Wang, Ly, Guo, and Kambhamettu} 2014
Citation
Citation
{Wang, Ly, Guo, and Kambhamettu} 2014
Citation
Citation
{Huang, Ardabilian, Wang, and Chen} 2012
Page 10: Improving Heterogeneous Face Recognition with Conditional

10 W. ZHANG ET AL.: IMPROVING HFR WITH CGAN

6 ConclusionIn this paper, we have presented a novel framework for 2D/2.5D heterogeneous face recog-nition together with depth face reconstruction. This approach combines the generative ca-pacity of conditional GAN and the discriminative feature extraction of deep CNN for cross-modality learning. The extensive experiments have convincingly evidenced that the proposedmethod successfully reconstructs realistic 2.5D from single 2D while being adaptive and suf-ficient for HFR. This architecture could hopefully be generalized to other heterogeneous FRtasks, such as visible light vs. near-infrared and 2.5D vs. forensic sketch, which provides aninteresting and promising prospect.

7 AcknowledgementThis work was supported in part by the French Research Agency, l’Agence Nationale deRecherche (ANR), through the Jemime project (N◦ contract ANR-13-CORD-0004-02), theBiofence project (N◦ ANR-13-INSE-0004-02) and the PUF 4D Vision project funded by thePartner University Foundation.

References[1] CASIA-3D FaceV1 database. http://biometrics.idealtest.org/, 2004.

[2] Akshay Asthana, Stefanos Zafeiriou, Shiyang Cheng, and Maja Pantic. Incrementalface alignment in the wild. In Proceedings of the IEEE Conference on Computer Visionand Pattern Recognition, pages 1859–1866, 2014.

[3] Aisha Azeem, Muhammad Sharif, Mudassar Raza, and Marryam Murtaza. A survey:Face recognition techniques under partial occlusion. Int. Arab J. Inf. Technol., 11(1):1–10, 2014.

[4] Volker Blanz and Thomas Vetter. Face recognition based on fitting a 3d morphablemodel. IEEE Transactions on pattern analysis and machine intelligence, 25(9):1063–1074, 2003.

[5] Kevin W Bowyer, Kyong Chang, and Patrick Flynn. A survey of approaches and chal-lenges in 3d and multi-modal 3d+ 2d face recognition. Computer vision and imageunderstanding, 101(1):1–15, 2006.

[6] A Roy Chowdhury, Rama Chellappa, Sandeep Krishnamurthy, and Tai Vo. 3d facereconstruction from video using a generic model. In Multimedia and Expo, 2002.ICME’02. Proceedings. 2002 IEEE International Conference on, volume 1, pages 449–452. IEEE, 2002.

[7] Ciprian Adrian Corneanu, Marc Oliu Simón, Jeffrey F Cohn, and Sergio Escalera Guer-rero. Survey on rgb, 3d, thermal, and multimodal approaches for facial expressionrecognition: History, trends, and affect-related applications. IEEE transactions on pat-tern analysis and machine intelligence, 38(8):1548–1568, 2016.

Page 11: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 11

[8] Changxing Ding and Dacheng Tao. A comprehensive survey on pose-invariant facerecognition. ACM Transactions on intelligent systems and technology (TIST), 7(3):37,2016.

[9] Douglas Fidaleo and Gérard Medioni. Model-assisted 3d face reconstruction fromvideo. In International Workshop on Analysis and Modeling of Faces and Gestures,pages 124–138. Springer, 2007.

[10] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley,Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. InAdvances in neural information processing systems, pages 2672–2680, 2014.

[11] Lie Gu and Takeo Kanade. 3d alignment of face in a single image. In 2006 IEEE Com-puter Society Conference on Computer Vision and Pattern Recognition (CVPR’06),volume 1, pages 1305–1312. IEEE, 2006.

[12] Geoffrey E Hinton and Ruslan R Salakhutdinov. Reducing the dimensionality of datawith neural networks. science, 313(5786):504–507, 2006.

[13] Di Huang, Mohsen Ardabilian, Yunhong Wang, and Liming Chen. Oriented gradientmaps based automatic asymmetric 3d-2d face recognition. In 2012 5th IAPR Interna-tional Conference on Biometrics (ICB), pages 125–131. IEEE, 2012.

[14] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image trans-lation with conditional adversarial networks. arXiv preprint arXiv:1611.07004, 2016.

[15] Yi Jin, Jiuwen Cao, Qiuqi Ruan, and Xueqiao Wang. Cross-modality 2d-3d face recog-nition via multiview smooth discriminant analysis based on elm. Journal of Electricaland Computer Engineering, 2014:21, 2014.

[16] Ioannis A Kakadiaris, George Toderici, Georgios Evangelopoulos, Georgios Passalis,Dat Chu, Xi Zhao, Shishir K Shah, and Theoharis Theoharis. 3d-2d face recognitionwith pose and illumination normalization. Computer Vision and Image Understanding,2016.

[17] Ira Kemelmacher-Shlizerman and Ronen Basri. 3d face reconstruction from a singleimage using a single reference face shape. IEEE Transactions on Pattern Analysis andMachine Intelligence, 33(2):394–405, 2011.

[18] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. arXiv preprintarXiv:1312.6114, 2013.

[19] Yuancheng Lee, Jiancong Chen, Ching-Wei Tseng, and Shang-Hong Lai. Accurate androbust face recognition from rgb-d images with a deep learning approach. In BMVC,2016.

[20] Feng Liu, Dan Zeng, Qijun Zhao, and Xiaoming Liu. Joint face alignment and 3d facereconstruction. In European Conference on Computer Vision, pages 545–560. Springer,2016.

[21] Iain Matthews, Jing Xiao, and Simon Baker. 2d vs. 3d deformable face models: Repre-sentational power, construction, and real-time fitting. International journal of computervision, 75(1):93–113, 2007.

Page 12: Improving Heterogeneous Face Recognition with Conditional

12 W. ZHANG ET AL.: IMPROVING HFR WITH CGAN

[22] Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. arXivpreprint arXiv:1411.1784, 2014.

[23] Unsang Park and Anil K Jain. 3d model-based face recognition in video. In Interna-tional Conference on Biometrics, pages 1085–1094. Springer, 2007.

[24] Omkar M Parkhi, Andrea Vedaldi, and Andrew Zisserman. Deep face recognition. InBMVC, volume 1, page 6, 2015.

[25] P Jonathon Phillips, Patrick J Flynn, Todd Scruggs, Kevin W Bowyer, Jin Chang, KevinHoffman, Joe Marques, Jaesik Min, and William Worek. Overview of the face recogni-tion grand challenge. In 2005 IEEE Computer Society Conference on Computer Visionand Pattern Recognition (CVPR’05), volume 1, pages 947–954. IEEE, 2005.

[26] Daniel Riccio and Jean-Luc Dugelay. Geometric invariants for 2d/3d face recognition.Pattern Recognition Letters, 28(14):1907–1914, 2007.

[27] Salah Rifai, Pascal Vincent, Xavier Muller, Xavier Glorot, and Yoshua Bengio. Con-tractive auto-encoders: Explicit invariance during feature extraction. In Proceedingsof the 28th international conference on machine learning (ICML-11), pages 833–840,2011.

[28] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networksfor biomedical image segmentation. In International Conference on Medical ImageComputing and Computer-Assisted Intervention, pages 234–241. Springer, 2015.

[29] Arman Savran, Nese Alyüz, Hamdi Dibeklioglu, Oya Çeliktutan, Berk Gökberk, Bü-lent Sankur, and Lale Akarun. Bosphorus database for 3d face analysis. In EuropeanWorkshop on Biometrics and Identity Management, pages 47–56. Springer, 2008.

[30] Yi Sun, Ding Liang, Xiaogang Wang, and Xiaoou Tang. Deepid3: Face recognitionwith very deep neural networks. arXiv preprint arXiv:1502.00873, 2015.

[31] Yaniv Taigman, Ming Yang, Marc’Aurelio Ranzato, and Lior Wolf. Deepface: Closingthe gap to human-level performance in face verification. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition, pages 1701–1708, 2014.

[32] Xiaoyang Tan, Songcan Chen, Zhi-Hua Zhou, and Fuyan Zhang. Face recognition froma single image per person: A survey. Pattern recognition, 39(9):1725–1745, 2006.

[33] George Toderici, Georgios Passalis, Stefanos Zafeiriou, Georgios Tzimiropoulos,Maria Petrou, Theoharis Theoharis, and Ioannis A Kakadiaris. Bidirectional relightingfor 3d-aided 2d face recognition. In Computer Vision and Pattern Recognition (CVPR),2010 IEEE Conference on, pages 2721–2728. IEEE, 2010.

[34] Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, and Pierre-AntoineManzagol. Stacked denoising autoencoders: Learning useful representations in a deepnetwork with a local denoising criterion. Journal of Machine Learning Research, 11(Dec):3371–3408, 2010.

[35] Xiaolong Wang, Vincent Ly, Rui Guo, and Chandra Kambhamettu. 2d-3d face recog-nition via restricted boltzmann machines. In Computer Vision Theory and Applications(VISAPP), 2014 International Conference on, volume 2, pages 574–580. IEEE, 2014.

Page 13: Improving Heterogeneous Face Recognition with Conditional

W. ZHANG ET AL.: IMPROVING HFR WITH CGAN 13

[36] Ziyan Wang, Ruogu Lin, Jiwen Lu, Jianjiang Feng, et al. Correlated and in-dividual multi-modal deep learning for rgb-d object recognition. arXiv preprintarXiv:1604.01655, 2016.

[37] Dong Yi, Zhen Lei, Shengcai Liao, and Stan Z Li. Learning face representation fromscratch. arXiv preprint arXiv:1411.7923, 2014.

[38] Lijun Yin, Xiaozhou Wei, Yi Sun, Jun Wang, and Matthew J Rosato. A 3d facialexpression database for facial behavior research. In 7th international conference onautomatic face and gesture recognition (FGR06), pages 211–216. IEEE, 2006.

[39] Wuming Zhang, Di Huang, Yunhong Wang, and Liming Chen. 3d aided face recog-nition across pose variations. In Chinese Conference on Biometric Recognition, pages58–66. Springer, 2012.

[40] Wenyi Zhao, Rama Chellappa, P Jonathon Phillips, and Azriel Rosenfeld. Face recog-nition: A literature survey. ACM computing surveys (CSUR), 35(4):399–458, 2003.

[41] Xi Zhao, Wuming Zhang, Georgios Evangelopoulos, Di Huang, Shishir K Shah, Yun-hong Wang, Ioannis A Kakadiaris, and Liming Chen. Benchmarking asymmetric 3d-2dface recognition systems. In Automatic Face and Gesture Recognition (FG), 2013 10thIEEE International Conference and Workshops on, pages 1–8. IEEE, 2013.