11
IEEE TRANSACTIONS ON MULTIMEDIA 1 Scalable Face Image Retrieval using Attribute-Enhanced Sparse Codewords Bor-Chun Chen, Yan-Ying Chen, Yin-Hsi Kuo, Winston H. Hsu Abstract—Photos with people (e.g., family, friends, celebrities, etc.) are the major interest of users. Thus, with the exponentially growing photos, large-scale content-based face image retrieval is an enabling technology for many emerging applications. In this work, we aim to utilize automatically detected human attributes that contain semantic cues of the face photos to improve content- based face retrieval by constructing semantic codewords for effi- cient large-scale face retrieval. By leveraging human attributes in a scalable and systematic framework, we propose two orthogonal methods named attribute-enhanced sparse coding and attribute- embedded inverted indexing to improve the face retrieval in the offline and online stages. We investigate the effectiveness of different attributes and vital factors essential for face retrieval. Experimenting on two public datasets, the results show that the proposed methods can achieve up to 43.5% relative improvement in MAP compared to the existing methods. Index Terms—Face image, human attributes, content-based image retrieval I. I NTRODUCTION D UE to the popularity of digital devices and the rise of social network/photo sharing services (e.g., Facebook, Flickr), there are largely growing consumer photos available in our life. Among all those photos, a big percentage of them are photos with human faces (estimated more than 60%). The importance and the sheer amount of human face photos make manipulations (e.g., search and mining) of large-scale human face images a really important research problem and enable many real world applications [1], [2]. Our goal in this paper is to address one of the important and challenging problems – large-scale content-based face image retrieval. Given a query face image, content-based face image retrieval tries to find similar face images from a large image database. It is an enabling technology for many applications including automatic face annotation [2], crime investigation [3], etc. Traditional methods for face image retrieval usually use low-level features to represent faces [2], [4], [5], but low-level Copyright (c) 2012 IEEE. Personal use of this material is permitted. However, permission to use this material for any other purposes must be obtained from the IEEE by sending a request to [email protected]. B.-C. Chen is with the Department of Computer Science and Information Engineering, National Taiwan University (e-mail: [email protected]). Y.-Y. Chen is with the Department of Computer Science and Information Engineering, National Taiwan University (e-mail: [email protected]). Y.-H. Kuo is with the Graduate Institute of Networking and Multimedia, National Taiwan University (e-mail: [email protected]). W. H. Hsu is with the Graduate Institute of Networking and Multimedia and the Department of Computer Science and Information Engineering, National Taiwan University, Taipei 10617, Taiwan (e-mail: [email protected]). Prof. Hsu is the contact person. This work was supported in part by grants from the National Science Council of Taiwan, under Contracts NSC 101-2628-E-002-027-MY2, and Excellent Research Projects of National Taiwan University, AE00-00-05. (a) Gender Hair Race (b) Similar in low-level appearance Different in attributes Fig. 1. (a) Because low-level features are lack of semantic meanings, face images of two different people might be close in the traditional low-level feature space. (b) By incorporating high-level human attributes (e.g., gender) into feature representations, we can provide better discriminability for face image retrieval. (Best seen in color) features are lack of semantic meanings and face images usually have high intra-class variance (e.g., expression, posing), so the retrieval results are unsatisfactory (cf. Figure 1 (a)). To tackle this problem, Wu et al. [4] propose to use identity based quantization and Chen et al. [5] propose to use identity- constrained sparse coding, but these methods might require clean training data and massive human annotations. In this work, we provide a new perspective on content- based face image retrieval by incorporating high-level human attributes into face image representation and index structure. As shown in Figure 1, face images of different people might be very close in the low-level feature space. By combining low-level features with high-level human attributes, we are able to find better feature representations and achieve better retrieval results. The similar idea is proposed in [6] using fisher vectors with attributes for large-scale image retrieval, but they use early fusion to combine the attribute scores. Also, they do not take advantages of human attributes because their target is general image retrieval. Human attributes (e.g., gender, race, hair style) are high- level semantic descriptions about a person. Some examples of human attributes can be found in Figure 2 (a). The recent work shows automatic attribute detection has adequate quality (more than 80% accuracy) [7] on many different human attributes. Using these human attributes, many researchers have achieved promising results in different applications such as face verification [7], face identification [8], keyword-based face image retrieval [9], and similar attribute search [10].

IEEE TRANSACTIONS ON MULTIMEDIA 1 ScalableFace Image ...winston/papers/chen12scalable.pdf · IEEE TRANSACTIONS ON MULTIMEDIA 1 ScalableFace Image Retrieval using Attribute-Enhanced

Embed Size (px)

Citation preview

IEEE TRANSACTIONS ON MULTIMEDIA 1

Scalable Face Image Retrieval usingAttribute-Enhanced Sparse Codewords

Bor-Chun Chen, Yan-Ying Chen, Yin-Hsi Kuo, Winston H. Hsu

Abstract—Photos with people (e.g., family, friends, celebrities,etc.) are the major interest of users. Thus, with the exponentiallygrowing photos, large-scale content-based face image retrieval isan enabling technology for many emerging applications. In thiswork, we aim to utilize automatically detected human attributesthat contain semantic cues of the face photos to improve content-based face retrieval by constructing semantic codewords for effi-cient large-scale face retrieval. By leveraging human attributes ina scalable and systematic framework, we propose two orthogonalmethods named attribute-enhanced sparse coding and attribute-embedded inverted indexing to improve the face retrieval inthe offline and online stages. We investigate the effectiveness ofdifferent attributes and vital factors essential for face retrieval.Experimenting on two public datasets, the results show that theproposed methods can achieve up to 43.5% relative improvementin MAP compared to the existing methods.

Index Terms—Face image, human attributes, content-basedimage retrieval

I. INTRODUCTION

DUE to the popularity of digital devices and the rise ofsocial network/photo sharing services (e.g., Facebook,

Flickr), there are largely growing consumer photos availablein our life. Among all those photos, a big percentage of themare photos with human faces (estimated more than 60%). Theimportance and the sheer amount of human face photos makemanipulations (e.g., search and mining) of large-scale humanface images a really important research problem and enablemany real world applications [1], [2].

Our goal in this paper is to address one of the important andchallenging problems – large-scale content-based face imageretrieval. Given a query face image, content-based face imageretrieval tries to find similar face images from a large imagedatabase. It is an enabling technology for many applicationsincluding automatic face annotation [2], crime investigation[3], etc.

Traditional methods for face image retrieval usually uselow-level features to represent faces [2], [4], [5], but low-level

Copyright (c) 2012 IEEE. Personal use of this material is permitted.However, permission to use this material for any other purposes must beobtained from the IEEE by sending a request to [email protected].

B.-C. Chen is with the Department of Computer Science andInformation Engineering, National Taiwan University (e-mail:[email protected]). Y.-Y. Chen is with the Department ofComputer Science and Information Engineering, National Taiwan University(e-mail: [email protected]). Y.-H. Kuo is with the GraduateInstitute of Networking and Multimedia, National Taiwan University (e-mail:[email protected]). W. H. Hsu is with the Graduate Institute ofNetworking and Multimedia and the Department of Computer Science andInformation Engineering, National Taiwan University, Taipei 10617, Taiwan(e-mail: [email protected]). Prof. Hsu is the contact person. Thiswork was supported in part by grants from the National Science Councilof Taiwan, under Contracts NSC 101-2628-E-002-027-MY2, and ExcellentResearch Projects of National Taiwan University, AE00-00-05.

(a)

Gender

Hair

Race

(b)

Similar in low-levelappearance

Different in attributes

Fig. 1. (a) Because low-level features are lack of semantic meanings, faceimages of two different people might be close in the traditional low-levelfeature space. (b) By incorporating high-level human attributes (e.g., gender)into feature representations, we can provide better discriminability for faceimage retrieval. (Best seen in color)

features are lack of semantic meanings and face images usuallyhave high intra-class variance (e.g., expression, posing), sothe retrieval results are unsatisfactory (cf. Figure 1 (a)). Totackle this problem, Wu et al. [4] propose to use identitybased quantization and Chen et al. [5] propose to use identity-constrained sparse coding, but these methods might requireclean training data and massive human annotations.

In this work, we provide a new perspective on content-based face image retrieval by incorporating high-level humanattributes into face image representation and index structure.As shown in Figure 1, face images of different people mightbe very close in the low-level feature space. By combininglow-level features with high-level human attributes, we areable to find better feature representations and achieve betterretrieval results. The similar idea is proposed in [6] using fishervectors with attributes for large-scale image retrieval, but theyuse early fusion to combine the attribute scores. Also, they donot take advantages of human attributes because their target isgeneral image retrieval.

Human attributes (e.g., gender, race, hair style) are high-level semantic descriptions about a person. Some examplesof human attributes can be found in Figure 2 (a). The recentwork shows automatic attribute detection has adequate quality(more than 80% accuracy) [7] on many different humanattributes. Using these human attributes, many researchershave achieved promising results in different applications suchas face verification [7], face identification [8], keyword-basedface image retrieval [9], and similar attribute search [10].

IEEE TRANSACTIONS ON MULTIMEDIA 2

These results indicate the power of the human attributes onface images. In Table I, we also show that human attributescan be helpful for identifying a person by the information-theoretic measures.

Although human attributes have been shown useful onapplications related to face images, it is non-trivial to applyit in content-based face image retrieval task due to severalreasons. First, human attributes only contain limited dimen-sions. When there are too many people in the dataset, it losesdiscriminability because certain people might have similarattributes. Second, human attributes are represented as a vectorof floating points. It does not work well with developing large-scale indexing methods, and therefore it suffers from slowresponse and scalability issue when the data size is huge.

To leverage promising human attributes automatically de-tected by attribute detectors for improving content-based faceimage retrieval, we propose two orthogonal methods namedattribute-enhanced sparse coding and attribute-embedded in-verted indexing. Attribute-enhanced sparse coding exploits theglobal structure of feature space and uses several importanthuman attributes combined with low-level features to constructsemantic codewords in the offline stage. On the other hand,attribute-embedded inverted indexing locally considers humanattributes of the designated query image in a binary signatureand provides efficient retrieval in the online stage. By incorpo-rating these two methods, we build a large-scale content-basedface image retrieval system by taking advantages of both low-level (appearance) features and high-level (facial) semantics.

In order to evaluate the performance of the proposed meth-ods, we conduct extensive experiments on two separate publicdatasets named LFW [11] and Pubfig [12]. These two datasetscontain faces taken in unconstrained environment and arereally challenging for content-based face image retrieval. Someexamples of the datasets can be found in Figure 6. During theexperiments, we show that the proposed methods can leveragethe context information from human attributes to achieverelative improvement up to 43.55% in mean average precisionon face retrieval task compared to the existing methods usinglocal binary pattern (LBP) [13] and sparse coding [5]. We alsoanalyze the effectiveness of different human attributes acrossdatasets and find informative human attributes.

To sum up, the contributions of this paper include:• We combine automatically detected high-level human

attributes and low-level features to construct semanticcodewords. To the best of our knowledge, this is thefirst proposal of such combination for content-based faceimage retrieval.

• To balance global representations in image collectionsand locally embedded facial characteristics, we proposetwo orthogonal methods to utilize automatically detectedhuman attributes to improve content-based face imageretrieval under a scalable framework.

• We conduct extensive experiments and demonstrate theperformances of the proposed methods on two separatepublic datasets and still ensure real time response.

• We further identify informative and generic human at-tributes for face image retrieval across different datasets.The selected descriptors are promising for other applica-

tions (e.g., face verification) as well.The rest of the paper is organized as follows. Section IIdiscusses related work. Section III describes our observationson the face image retrieval problem and the promising utilitiesof human attributes. Section IV introduces the proposed meth-ods including attribute-enhanced sparse coding and attribute-embedded inverted indexing. Section V gives the experimentalresults, and section VI concludes this paper.

II. RELATED WORK

This work is closely related to several different researchtopics, including content-based image retrieval (CBIR), humanattribute detection, and content-based face image retrieval.

Traditional CBIR techniques use image content like color,texture and gradient to represent images. To deal with large-scale data, mainly two kinds of indexing systems are used.Many studies have leveraged inverted indexing [14] or hash-based indexing [15] combined with bag-of-word model (BoW)[16] and local features like SIFT [17], to achieve efficientsimilarity search. Although these methods can achieve highprecision on rigid object retrieval, they suffer from low re-call problem due to the semantic gap [18]. Recently, someresearchers have focused on bridging the semantic gap byfinding semantic image representations to improve the CBIRperformance. [19] and [20] propose to use extra textual in-formation to construct semantic codewords; [21] uses classlabels for semantic hashing. The idea of this work is similarto the aforementioned methods, but instead of using extrainformation that might require intensive human annotations(and tagging), we try to exploit automatically detected humanattributes to construct semantic codewords for the face imageretrieval task.

Automatically detected human attributes have been shownpromising in different applications recently. Kumar et al.propose a learning framework to automatically find describablevisual attributes [7]. Using automatically detected humanattributes, they achieve excellent performance on keyword-based face image retrieval and face verification. Siddiquie et al.[9] further extend the framework to deal with multi-attributequeries for keyword-based face image retrieval. Scheirer etal. [8] propose a Bayesian network approach to utilize thehuman attributes for face identification. To further improvethe quality of attributes, Parikh et al. propose to use relativeattributes [22] and Scheirer et al. propose multi-attribute space[10] to normalize the confidence scores from different attributedetectors for similar attribute search. The works demonstratethe emerging opportunities for the human attributes but are notexploited to generate more semantic (scalable) codewords. Al-though these works achieve salient performance on keyword-based face image retrieval and face recognition, we proposeto exploit effective ways to combine low-level features andautomatically detected facial attributes for scalable face imageretrieval. To the best of our knowledge, very few works aimto deal with this problem.

Due to the rise of photo sharing/social network services,there rises the strong needs for large-scale content-basedface image retrieval. Content-based face image retrieval isclosely related to face recognition problems but they focus on

IEEE TRANSACTIONS ON MULTIMEDIA 3

TABLE IENTROPY AND MUTUAL INFORMATION COMPUTED FROM TWO DIFFERENTDATASETS. X IS A RANDOM VARIABLE FOR THE IDENTITY OF A PERSON.

Y IS THE ATTRIBUTE. THE CONDITIONAL ENTROPY, GIVEN THEATTRIBUTE (E.G., GENDER), DROPS. IT SUGGESTS THAT USING HUMAN

ATTRIBUTES CAN HELP IDENTIFY A PERSON.

Dataset H(X) H(X|Y ) I(X;Y )

LFW 11.21 10.45 0.77Pubfig 5.43 4.46 0.97

finding suitable feature representations for scalable indexingsystems. Because face recognition usually requires substantialcomputation cost for dealing with high dimensional featuresand generating explicit classification models, it is non-trivial todirectly apply it to face retrieval tasks. Meanwhile, the photoquality in consumer photos is more diverse and poses morevisual variances. Wu et al. [4] propose a face retrieval frame-work using component-based local features with identity-basedquantization to deal with scalability issues. To compensatethe quantization loss, they further propose to use a state-of-the-art features [23] with principal component analysis for re-ranking. Wang et. al. [2] propose an automatic face annotationframework based on content-based face image retrieval. Intheir framework, they adopt GIST [24] feature with localitysensitive hashing [15] for face image retrieval. Chen et al. [5]propose to use component-based local binary pattern (LBP)[13], a well known feature for face recognition, combinedwith sparse coding and partial identity information to constructsemantic codewords for content-based face image retrieval.

Although images naturally have very high dimensional rep-resentations, those within the same class usually lie on a lowdimensional subspace. Sparse coding can exploit the semanticsof the data and achieve promising results in many differentapplications such as image classification and face recognition.Raina et al. propose a machine learning framework usingunlabeled data with sparse coding for classification tasks [25],Yang et al. [26] apply the framework on SIFT descriptors alongwith spatial pyramid matching [27] and maximum pooling toimprove classification results. Wright et. al. [28] propose to usesparse representation for face recognition and achieve state-of-the-art performance.

Taking advantages of the effectiveness and simplicity ofLBP feature with the superior characteristics of sparse codingon face images, we adopt a similar approach as Chen et al.used in [5] – using component-based LBP combined withsparse coding to construct sparse codewords for efficientcontent-based face image retrieval. However, instead of usingidentity information that might need manual annotations, wefocus on utilizing automatically detected human attributes toconstruct semantic-aware sparse codewords using attribute-enhanced sparse coding. In addition, we propose another or-thogonal approach to further leverage attribute information byconstructing attribute-embedded inverted index in the onlineranking stage. Note that the proposed methods can be easilycombined with the method proposed in [5] to take advantageof both identity information and automatically detected humanattributes. Also, low-level feature (i.e., LBP) can be replacedby other features such as T3HS2 descriptor used in [4].

(b)(a)

Male: -0.35Brown Hair: 1.44

Youth: 0.40Eyeglasses: -2.06

Smiling: 0.66...

Fig. 2. (a) A face image contains rich context information (hair color, skincolor, race, gender, etc.). Using automatic human attribute detection, we candescribe them in high-level semantics; for example, Male: -0.35 suggests theperson is unlikely a male, and Brown Hair: 1.44 implies the hair color tends tobe brown. (b) The same image after preprocessing steps in prior works for faceimage retrieval or recognition. They normalize the position and illuminationdifferences between the faces and exclude background contents. Such commonapproaches sacrifice the important context information. Using automaticallydetected human attributes can compensate the information loss.

III. OBSERVATIONS

When dealing with face images, prior works [2], [4], [5]usually crop only the facial region and normalize the faceinto the same position and illumination to reduce intra-classvariance caused by poses and lighting variations. Doing thesepreprocessing steps, they ignore the rich semantic cues fora designated face such as skin color, gender, hair style. Toillustrate, Figure 2 shows the face image before and afterthe common preprocessing steps. After preprocessing steps,the information loss causes difficulty in identifying attributes(e.g., gender) of the face. In [7], the authors conducted humanexperiments to support similar points. When using a croppedversion of face images, the face verification performance willdrop comparing with using the original uncropped version.Interestingly, their experiments also show that human canachieve salient verification performance using only the sur-rounding context of face images. The experiments suggest thatthe surrounding context indeed contains important informationfor identifying a person. Therefore, we propose to use automat-ically detected human attributes to compensate the informationloss.

Given a face image, let X be a random variable for theidentity of a person, and Y is the attribute (e.g., gender).In information-theoretic perspective, knowing attributes canreduce the entropy for identifying a person and the informationgain can be computed as,

I(X;Y ) = H(X)−H(X|Y ), (1)

where H(X) denotes the Shannon entropy of X , which isused to measure the uncertainty of the random variable X .H(X|Y ) is the conditional entropy of X given Y and showsthe uncertainty of X after knowing the value of Y . Intuitively,the larger mutual information indicates more help coming fromY for predicting X . Table I shows the entropy and mutualinformation computed from two different public datasets usingonly gender as the human attribute. The detailed statistics ofthe datasets can be found in Table II. The probability of Xis computed by the frequency of the person in the dataset.Genders of the people are manually labeled. We only considergender in Y for simplicity. As a result, we gain up to 0.97 bitinformation, that is, considering the gender attribute allowsus to skip nearly half of the database if the database contains

IEEE TRANSACTIONS ON MULTIMEDIA 4

Query image

Preprocessing

Face detection

Facial landmark detection

Face alignment

Aligned face

Local patches Patch-level

LBP features

Attribute enhanced sparse coding

(Sec. IV-B)

Attribute embedded inverted indexing

(Sec. IV-C)

Attribute detection

Ranking results

Large-scale face Image database

Patch-levelsparse codewords

Male

White

Asia

n

Black

Gla

sses

Lipst

ick

Fig. 3. The proposed system framework. Both query and database images will go through the same procedures including face detection, facial landmarkdetection, face alignment, attribute detection, and LBP feature extraction. Attribute-enhanced sparse coding is used to find sparse codewords of databaseimages globally in the offline stage. Codewords of the query image are combined locally with binary attribute signature to traverse the attribute-embeddedinverted index in the online stage and derive real-time ranking results over database images.

50% females and 50% males. Hence we hypothesize that usinghuman attributes can help the face retrieval task.

IV. PROPOSED METHOD

In this section, we first describe the overview of our scalablecontent-based face image retrieval system, and then we explainthe proposed methods: attribute-enhanced sparse coding andattribute-embedded inverted indexing in details. Note thatthe human attributes mentioned in the coming sections areautomatically detected using the method described in [7].

A. System overview

For every image in the database, we first apply Viola-Jonesface detector [29] to find the locations of faces. We then use theframework proposed in [7] to find 73 different attribute scores.Active shape model [30] is applied to locate 68 differentfacial landmarks on the image. Using these facial landmarks,we apply barycentric coordinate based mapping process toalign every face with the face mean shape [3]. For eachdetected facial component, we will extract 7×5 grids, whereeach grid is a square patch [4]. In total we have 175 gridsfrom five components including two eyes, nose tip, and twomouth corners. on the aligned image using similar methodsproposed in [4]. From each grid, we extract an image patchand compute a 59-dimensional uniform LBP feature descriptoras our local feature. After obtaining local feature descriptors,we quantize every descriptor into codewords using attribute-enhanced sparse coding described in section IV-B. Attribute-embedded inverted index described in section IV-C is thenbuilt for efficient retrieval. When a query image arrives, it willgo through the same procedure to obtain sparse codewordsand human attributes, and use these codewords with binaryattribute signature to retrieve images in the index system.Figure 3 illustrates the overview of our system.

B. Attribute-enhanced sparse coding (ASC)

In this section, we first describe how to use sparse coding forface image retrieval. We then describe details of the proposedattribute-enhanced sparse coding. Note that in the followingsections, we apply the same procedures to all patches in asingle image to find different codewords and combine all thesecodewords together to represent the image.

1) Sparse coding for face image retrieval (SC): Usingsparse coding for face image retrieval, we solve the followingoptimization problem:

minD,V

n∑i=1

||x(i) −Dv(i)||2

2 + λ||v(i)||1

subject to ||D∗j ||22 = 1,∀j(2)

where x(i) is the original features extracted from a patch offace image i, D ∈ Rd×K is a to-be-learned dictionary containsK centroids with d dimensions. V = [v(1), v(2), . . . , v(n)] isthe sparse representation of the image patches. The constrainton each column of D (D∗j) is to keep D from becomingarbitrarily large. Using sparse coding, a feature is a linearcombination of the column vectors of the dictionary. [31]provides an efficient online algorithm for solving the aboveproblem.

Note that the Equation (2) actually contains two parts:dictionary learning (find D) and sparse feature encoding (findV ). In [32], Coates et. al. found that using randomly sampledimage patches as dictionary can achieve similar performanceas that by using learned dictionary (< 2.7% relative improve-ment in their experiments) if the sampled patches provide a setof overcomplete basis that can represent input data. Becauselearning dictionary with a large vocabulary is time-consuming(training 175 codebooks with 1600 dimension takes morethan two weeks to finish), we can just use randomly sampledimage patches as our dictionary and skip the time-consumingdictionary learning step by fixing D in the Equation (2) anddirectly solve V . When D is fixed, the problem becomes aL1 regularized least square problem, and can be efficientlysolved using LARS algorithm [33]. After finding v(i) for eachimage patch, we consider nonzero entries as codewords ofimage i and use them for inverted indexing. Note that we applythe above process to 175 different spatial grids separately, socodewords from different grids will never match. Accordingly,we can encode the important spatial information of faces intosparse coding. The choice of K is investigated in Section V-A.We use K = 1600 in the experiments, so the final vocabularysize of the index system will be 175× 1600 = 280, 000.

2) Attribute-enhanced sparse coding (ASC): In order toconsider human attributes in the sparse representation, we first

IEEE TRANSACTIONS ON MULTIMEDIA 5

(a) SC (b) ASC-D (c) ASC-W

Fig. 4. Comparison between attribute enhancing coding methods: SC, ASC-Dand ASC-W. Colors denote the attribute of the images (patches) and the shapes(e.g., triangle, circle) indicate the identity. (a) Because SC only considers low-level features, the images with the same attribute might be assigned to differentdictionary centroids. (b) Using dictionary selection to force images with thesame attribute being assigned to similar dictionary centroids. However, ASC-D uses attributes with hard assignment, when the detected attribute is wrong,the result will be undesired. In (b), blue triangle in the bottom-right corneris assigned to totally different codewords from the other two triangles. (c)Using attribute weights instead of inflexible dictionary selection, even whendetected attributes are not perfect, it might still be able to retrieve correctimages if appearance of the images is smiliar enough, because we use softweights instead of hard assignment. (Best seen in color)

propose to use dictionary selection (ASC-D) to force imageswith different attribute values to contain different codewords.For a single human attribute, as shown in Figure 4 (b), wedivide dictionary centroids into two different subsets, imageswith positive attribute scores (blue ones in Figure 4) willuse one of the subset and images with negative attributescores will use the other. For example, if an image has apositive male attribute score, it will use the first half of thedictionary centroids. If it has a negative male attribute score,it will use the second half of the dictionary centroids. Bydoing these, images with different attributes will surely havedifferent codewords. For the cases of multiple attributes, wedivide the sparse representation into multiple segments basedon the number of attributes, and each segment of sparserepresentation is generated depending on single attribute.

The above goal can be achieved by solving the followingoptimization problem modified from Equation (2):

minV

n∑i=1

||x(i) −Dv(i)||22 + λ||z(i) ◦ v(i)||1

z(i)j =

∞, if (1) j ≥ bK2 c and fa(i) ≥ 0

(2) j < bK2 c and fa(i) < 0

1, otherwise,

(3)

where “◦” denotes the pairwise multiplication between twovectors, fa(i) is the attribute score for ith image, and z(i) isa mask vector for deciding which codewords are allowed tobe used by image i. By using the mask vector z(i), it forcesthe sparse representation v(i)j to be zero if z(i)j is ∞ becauseany other values in these dimensions will cause the objectivefunction to become infinity. The final sparse representationv(i) can be found by solving a L1 regularized least squareproblem and only considering the dimensions where z(i)j = 1.Figure 4 (b) illustrates the above method. When an inputimage contains positive attribute value, the mask vector willbe [1, 1, . . . , 1,∞,∞, . . . ,∞]T , so only the first half of thesparse representation are likely to contain non-zero value forthe image patches, and vice versa. To simplify the notation,the above equation only considers single attribute. For multipleattributes, we can simply define z(i) depending on multipleattributes. For example, if there are two attributes and image

i contains positive scores for both attributes, z(i) will be-come [1, . . . , 1,∞, . . . ,∞, 1, . . . , 1,∞, . . . ,∞]T . AlthoughEquation (3) can successfully encode the human attributesinto the sparse representation, it has two main problems.First, it is not robust to possible attribute detection errors.Since our human attributes are automatically detected, theyare not error-free. When the detected attribute is wrong, itwill force images of the same person to be associated withtotally different codewords (cf. Figure 4 (b)). Second, it onlyencodes the human attributes as binary indicators but weactually have relative confidence scores. Therefore, we furtherpropose to integrate the relative scores of human attributesto relax Equation (3) into a soft weighted version (ASC-W)by defining z(i) in Equation (3) based on the attribute scoresof images. Motivated by [34], we first assign a half of thedictionary centroids to have +1 attribute score and use themto represent images with the positive attribute; the other halfof the dictionary centroids are assigned with −1 to representimages with the negative attribute. After the assignment, wecan use the distance between attribute scores of the imageand the attribute scores assigned to the dictionary centroids asthe weights for selecting codewords. Because the weights aredecided by attribute scores, two images with similar attributescores will have similar weight vector, and therefore havea higher chance to be assigned with similar codewords andresult in similar sparse representations. Figure 4 (c) illustratesthe above method, images with similar attributes will beassigned with similar centroids, but images with erroneousattributes might still be able to retrieve correct images if theiroriginal features are similar (cf. Figure 4 (c) bottom-rightblue triangle). In details, we first define an attribute vectora ∈ {1,−1}K , where aj contains the attribute scores of thejth centroid as follows,

aj =

{+ 1, if j ≥ bK2 c− 1, otherwise,

(4)

then we change the z(i) in Equation (3) to become

z(i)j = exp(

d(fa(i), aj)

σ), (5)

where d(fa(i), aj) is the distance between the attribute scoreof the ith image patch and that of the jth dictionary centroid,and σ is used to adjust the decaying weights. To solve theabove problem, we use the modified version of the LARSalgorithm [33] by adjusting the weights according to z(i)j .C. Attribute Embedded Inverted Indexing (AEI)

The methods described in Section IV-B aim to constructcodewords enhanced by human attributes. In this section wedescribe the second method that can utilize human attributesby adjusting the inverted index structure.

1) Image ranking and inverted indexing: For each image,after computing the sparse representation using the methoddescribed in Section IV-B, we can use codeword set c(i) torepresent it by taking non-zero entries in the sparse represen-tation as codewords. The similarity between two images arethen computed as follows,

S(i, j) = ||c(i) ∩ c(j)|| (6)The image ranking according to this similarity score can beefficiently found using inverted index structure.

IEEE TRANSACTIONS ON MULTIMEDIA 6

Attribute EmbeddedInverted Index

1110100101 01111

11101

0110001101

1100000000

10011

10110

01001

01100

01111

0001010000

11011

10010

Sparse codewords

Query Image

01101attribute signature

Fig. 5. Illustration of attribute-embedded inverted indexing. By consideringbinary attribute signature in the index system, we can skip images withlarge hamming distance in attribute hamming space and improve retrievalperformance when traversing the codeword index in the online stage.

(a) (b)

Fig. 6. Sample images from LFW (a) and Pubfig (b). Images in the samecolumn are of the same person. These images contain large variances inexpression, pose and illuminance which are very challenging for face retrieval.

2) Attribute-embedded inverted indexing: To embed at-tribute information into index structure, for each image, inaddition to sparse codewords c(i) computed from the facialappearance, we use a db dimension binary signature to repre-sent its human attribute, b(i):

b(i)j =

{1 if f (i)a (j) > 0

0 otherwise,(7)

The similarity score is then modified into,

S(i, j) =

{||c(i) ∩ c(j)|| if h(b(i), b(j)) ≤ T0 otherwise,

(8)

where h(i, j) denotes hamming distance between i and j,and T is a fixed threshold such that 0 ≤ T ≤ db. Asshown in Figure 5, attribute-embedded inverted index is builtusing the original codewords and the binary attribute signa-tures associated with all database images. The image rankingaccording to Equation (8) can still be efficiently computedusing inverted index by simply doing a XOR operation tocheck the hamming distance before updating the similarityscores. As mentioned in [35], since XOR operation is fasterthan updating scores, by skipping images with high hammingdistance in attribute hamming space, the overall retrieval timesignificantly decreases. Note that storage of the inverted indexcan be further compressed using many different techniques ininformation retrieval [14].

V. EXPERIMENTS

A. Experimental Setting

1) Datasets: We use two different public datasets (LFW[11] and Pubfig [12]) for the following experiments. LFWdataset contains 13,233 face images among 5,749 people, and12 people have more than 50 images. We take 10 images fromeach of these 12 people as our query set (120 images) and allother images as our database (13,113 images). In Pubfig [12],we take 100 images from 43 people as our database images(4,300 images) and 10 images each from those 43 people asour query set (430 images). Example images from these twodatasets can be found in Figure 6. The facial attribute scores ofPubfig and LFW are provided by [12], which use pre-trainedfacial attribute detectors to measure 73 attribute scores. Notethat the 73 attribute scores for these two datasets are alsopublicly available [12].

2) Compared algorithms: We use several different base-lines to compare with the proposed methods including twostate-of-the-art face recognition features. The methods aredescribed as follows: (1) LBP : concatenated 59-dimensionuniform LBP [13] features computed from 175 local patchesdescribed in section IV-A, i.e., totally 10325 dimensions; (2)ATTR : 73 dimensional human attributes computed by themethod described in [7]; (3) SC : the sparse representationcomputed from LBP features using 1600 random samples asdictionary centroids combined with inverted indexing. Similarmethods are used in [5]; (4) SC-original: similar to the featureextracted by (3), but we directly use the weight of sparserepresentation with linear search instead of using invertedindex. (5) ASC-D : attribute-enhanced sparse representationwith dictionary selection using the method described in Sec-tion IV-B2. (6) ASC-W : attribute-enhanced sparse represen-tation with attribute weights using the method described inSection IV-B2. (7) AEI : attribute-embedded inverted indexingdescribed in Section IV-C.

Note that although these datasets are widely used by manyresearches, most literatures use these datasets for research onface verification. In [4], they also use LFW dataset for faceimage retrieval. However, in their experiments, they need a(name) labeled training set besides the original dataset, sowe are unable to reproduce their results. Also we notice thatthe results highly depend on the selection of query images-some of the query images can achieve more than 90%average precision (AP), some other query images can onlyachieve less than 5% AP. Since they do not release theirexperimental settings, we cannot directly compare with theirwork. Therefore we set some state-of-the-art (conventionalsparse coding and LBP) baselines in our own work.

Throughout the experiments we use mean average precision(MAP) and precision at K (P@K) as our performance metric.Note that for SC, ASC-D and ASC-W, we use random samplesas dictionary centroids. We aim at a scalable framework forexponentially growing large-scale (and unseen) face photosrather than optimizing a specific benchmark. Therefore, inorder to make sure that the performance does not overfit thespecific dataset and our system can be generalized to otherdatasets, for experiments on LFW dataset, we use samples

IEEE TRANSACTIONS ON MULTIMEDIA 7

from Pubfig dataset and vice versa. The two datasets aredesigned with no overlapping in people, so there will be noimages of the same person in dictionary centroids and databaseimages.

3) Parameter settings: In SC and ASC we need to decideparameter λ and dictionary size K. In order to decide theparameters, we run sensitive test and find out that the behavioris similar on both datasets. In the experiments, we run differentdictionary sizes from 100 to 3200 and λ from 10−6 to 10−1.When λ is too large (λ = 10−1), almost all entries in sparserepresentation become zero, so the performance will drop.When λ and K are set properly (λ = [10−6, 10−2],K =[100, 3200]), the performance of SC is stable regardless theparameters on both datasets. For fair comparisons, we setK = 1600 on SC and ASC for both datasets. For SC, we usethe best λ according to LFW for experiments on Pubfig andvice versa. Because the parameters are not sensitive to datasets,we only use one of the setting (the parameters decided bysensitivity test on Pubfig) for the other experiments. In ASC-W, we also need to decide the weight decay parameter σ. Wefix K = 1600 and λ = 10−2, run experiments using ASC-Wwith Male attribute and σ from 10 to 200 and set σ = 120 forall following experiments.

B. Baseline performance

Table II shows the statistics of two datasets and the per-formance of four different baseline methods. In this work,we would like to highlight what improvements we can bringin as exploiting face attributes for semantic-rich sparse code-word representations. That is why we need to compare withLBP (one of the state-of-the-art low-level features), humanattributes alone (ATTR), and the conventional sparse codingmethod. The setting in LBP (175 grids × 59 dimensions)reaches the best performance in our experiments and it per-forms better than the original setting (49 grids × 59 dimen-sions) in [13] because it is more robust to pose variations. Inaddition, the 73 attributes have covered most of the generalattributes discussed in the related work (e.g., [7]), which isthe state-of-the-art of multiple facial attribute detection. ATTRobtains low P@10 in both datasets because it only has 73dimensions and has very limited discriminability. In addi-tion, ATTR obtains slightly higher MAP (15.1%) comparedto SC-based approach (14.7%). However, SC has superiorP@10 (49.0%) compared to ATTR (39.7%) because sparsecodewords preserve the distinctive face traits of a specificperson and benefit retrieving the faces with similar visualappearances. Therefore, we suggest that attribute-enhancedsparse codewords would further improve the accuracy ofcontent-based face image retrieval. The experiments will bediscussed in Section V-C 1. The results of SC is consistentwith the findings in [5]. Note that the performance in LFWdataset is slightly different from the number reported in [5].There are two reasons for the performance difference. First, werandomly select different queries for evaluation, therefore, theoverall MAP is slightly different. Second, the result reported

1We also combine SC and ATTR using late fusion technique. The resultsof late fusion can achieve up to 17.3% relative improvement over SC.

in [5] is based on the parameters discovered within the singledataset; while the result in this paper is using the parametersdetermined by cross-validation manner over the two datasets.

We also compare with the SC-original to illustrate the su-periority of inverted index. We find that SC-original for linearsearch takes average 1.73 seconds to retrieval LFW datasetwith 13K images while using SC for inverted index onlytakes about 0.03 seconds; the speedup is 57.7. Interestingly,we also find out that binarizing 2 SC actually improves theperformance because it increases the discriminability betweennon-zero entries and zero entries of the feature vector. If wedo not binarize the sparse representation, the number in eachnon-zero dimension is too small as comparing with the highdimension of the feature vector. Also, it is worth noting thatthe performance on Pubfig dataset is better than LFW dataset,it is also consistent with the entropy measure in Table I,because higher entropy indicates the task is harder and requiresadditional information.

C. Experiments on attribute-enhanced sparse coding

1) Performance of single attribute: Table III shows theMAP of attribute-enhanced sparse coding based on singleattribute. We have computed the performance of ASC basedon 73 different attributes on both datasets, due to the spacelimitation, we only show the results of top 10 and bottom10 attributes ranked by ASC-W. We find that based ononly single attribute, we can achieve up to 12.2% relativeimprovement in Pubfig dataset and 16.2% in LFW datasetusing ASC-W. For some attributes, ASC-D performs worsethan SC, because ASC-D is highly correlated with attributedetection accuracy. ASC-W performs better than ASC-D inmost attributes, because ASC-W can take advantage of therelative attribute scores and is thus more robust to attributedetection error.

We also notice that using certain attributes (smiling, frown-ing, harsh lighting, etc.) will decrease the performance inboth datasets. It is probably because these attributes are notcorrelated with the identity of the person. Informative humanattributes across both datasets are also similar. Six out of thetop ten attributes are overlapped in two datasets, while 15 outof the top 20 attributes are overlapped. To further demonstrateour findings, we compute the Spearman’s rank correlationcoefficient (ρ)3 of two ranking lists and the coefficient is0.89. It suggests the rankings from the two datasets are highlycorrelated, therefore the informative attributes are consistentacross human photos.

2) Informative attributes discovery and performance ofmultiple attributes: Using attributes ranked by ASC-W acrosstwo datasets, we are able to identify informative attributes.Here we categorize top 20 overlapped attributes into fivedifferent categories. Including gender related attributes (G):Male, Wearing Earrings, Wearing Lipstick, Attractive Woman;

2For sparse coding, only the important codewords will be assigned withnon-zero value. Therefore, we assign these non-zero elements with 1 and theothers with 0.

3Spearman’s rank correlation coefficient is a measure used in statistics tocompute the correlation between the ranked variables, it range from -1 (perfectnegatively correlated) to 1 (perfect positively correlated).

IEEE TRANSACTIONS ON MULTIMEDIA 8

TABLE IIDATASET STATISTICS AND BASELINE PERFORMANCE. OVERALL PERFORMANCE IS WORSE ON LFW DATASET THAN ON PUBFIG, BECAUSE LFW

CONTAINS MORE IMAGES AND MORE DIFFERENT PEOPLE. ATTR HAS THE HIGHEST MAP IN PUBFIG BUT LOWEST MAP IN LFW BECAUSE ATTR LOSESDISCRIMINABILITY WHEN THERE ARE TOO MANY PEOPLE IN THE DATASET. ATTR PERFORMS THE WORST IN TERMS OF P@10 BECAUSE IT ONLY HASLIMITED DIMENSIONS. ALSO NOTE THAT LBP AND ATTR REQUIRE LINEAR SEARCH AND ARE NOT SCALABLE IN LARGE-SCALE DATASET. HOWEVER,

WE SHOW HOW ATTRIBUTE-ENHANCED SPARSE CODEWORDS CAN HELP LARGE-SCALE FACE RETRIEVAL.

LFW Pubfig# of people 5749 43

Database size 13113 4300# of queries 120 430Performance MAP P@10 Time (s) MAP P@10 Time (s)

LBP 11.9% 49.6% 1.01 11.6% 47.4% 0.38ATTR 11.6% 37.8% 0.04 15.1% 39.7% 0.01

SC 13.0% 46.8% 0.03 14.7% 49.0% 0.01SC-original 10.0% 39.3% 1.73 10.9% 39.6% 0.59

TABLE IIIMAP OF ATTRIBUTE-ENHANCED SPARSE CODING USING A SINGLE ATTRIBUTE. SOME ATTRIBUTES HAVE A NEGATIVE EFFECT ON PERFORMANCE

BECAUSE THOSE ATTRIBUTES ARE NOT CORRELATED WITH THE IDENTITY. ASC-W PERFORMS BETTER IN MOST ATTRIBUTES BECAUSE IT CONSIDERSTHE RELATIVE SCORES OF ATTRIBUTES AND IS ROBUST TO POSSIBLE ATTRIBUTE DETECTION ERROR. INTERESTINGLY, TOP RANKED AND BOTTOM

RANKED ATTRIBUTES ARE SIMILAR ACROSS THE TWO DATASETS. SPEARMAN’S RANK CORRELATION COEFFICIENT BETWEEN TWO RANKING LIST USINGASC-W IS 0.89; IT INDICATES THAT ATTRIBUTE EFFECTIVENESS RANKED BY ASC-W IS VERY SIMILAR ACROSS TWO DATASETS.

LFW PubfigTop 10 ASC-D ASC-W Top 10 ASC-D ASC-WSenior 12.5% 15.1% Male 16.1% 16.5%White 13.0% 14.9% Bushy Eyebrows 14.5% 16.2%Bald 11.4% 14.8% Receding Hairline 14.3% 16.2%

Black Hair 13.2% 14.6% No Eyewear 15.7% 16.1%Gray Hair 10.1% 14.6% Black Hair 13.4% 16.1%

Receding Hairline 12.0% 14.6% Eyeglasses 15.9% 16.0%Bushy Eyebrows 10.5% 14.5% Wearing Lipstick 15.8% 16.0%

Youth 12.8% 14.4% Youth 13.2% 15.9%No Eyewear 12.7% 14.4% Blond Hair 14.4% 15.9%

Wearing Earrings 13.1% 14.4% Bald 13.6% 15.9%Bottom 10 ASC-D ASC-W Bottom 10 ASC-D ASC-W

Teeth Not Visible 10.1% 13.0% Harsh Lighting 10.8% 13.7%Flash 11.1% 12.9% Smiling 11.8% 13.7%

Harsh Lighting 9.9% 12.8% Frowning 11.8% 13.7%Mouth Closed 9.0% 12.5% Soft Lighting 8.9% 13.6%

Frowning 10.7% 12.4% Mouth Closed 10.9% 13.6%Smiling 10.5% 12.4% Fully Visible Forehead 11.3% 13.5%

Rosy Cheeks 10.3% 12.2% Rosy Cheeks 10.7% 12.9%Flushed Face 10.9% 12.1% Flushed Face 11.0% 12.4%

Pale Skin 8.6% 11.8% Pale Skin 9.1% 12.0%Color Photo 13.2% 11.0% Color Photo 14.2% 10.9%

TABLE IVPERFORMANCE OF ASC-W USING MULTIPLE ATTRIBUTES. WE CAN

ACHIEVE UP TO 27.9% RELATIVE IMPROVEMENT IN LFW AND 26.9% INPUBFIG USING TWO DIFFERENT ATTRIBUTES AGAINST THE BASELINES.USING THREE DIFFERENT ATTRIBUTES, WE CAN ACHIEVE UP TO 41.0%

RELATIVE IMPROVEMENT IN LFW AND 36.9% IN PUBFIG.

Attributes LFW PubfigA (Senior)/H (Gray Hair) 16.6% 17.3%G (Male)/P (No Eyewear) 15.8% 18.7%R (White)/A (Senior)/P (Bald) 18.3% 18.3%G (Male)/H (Black Hair)/ 17.3% 20.2%P (No Eyewear)

hair colors (H): Blond Hair, Black Hair, Gray Hair; races(R): White, Asian, Indian; ages (A): Youth, Senior; andpersonal features (P): Receding Hairline, Bald, No Beard,Mustache, No Eyewear, Eyeglasses, Bushy Eyebrows, DoubleChin. Combining these attributes using ASC-W, we can furtherachieve better performance. Table IV shows some exampleresults of ASC-W which combines several attributes. We canachieve higher performance by using more than one attribute.By choosing two attributes from the above categories, the

relative improvement can be up to 27.9% in LFW dataset and26.9% in Pubfig dataset. When using three different attributes,we can achieve up to 41.0% and 36.9% relative improvementrespectively.

Note that when more attributes are used, each attribute-enhanced patch is represented by less codewords if the numberof feature dimensions is fixed. However, in our experiments,if we use up to 16 attributes, each attribute will be representedby a dictionary with size of 100 (for generating a 1,600dimensional feature), which is still an over-complete basiscompared to the dimension of the LBP feature (which is59 dimensions) and therefore still has enough discriminativepower for sparse feature representation. As suggested in [32],when using sparse coding as an encoder, the main contributionof the dictionary is to provide a over-complete basis. For moreattributes, we need a larger dictionary with more codewords torepresent the images; this is a trade-off between performanceand online retrieval speed.

In our experiments, we show that the proposed methodscan achieve salient performance (cf. Table IV) using only

IEEE TRANSACTIONS ON MULTIMEDIA 9

!"#$

!"##$

!"#%$

!"#&$

!"#'$

!"#($

!"#)$

!"#*$

!"#+$

!"%$ !"&$ !"'$ !"($ !")$ !"*$

!"#!

$%&'(!

,-.$/$012'!$

,-.$

Fig. 7. The results of attribute-embedded inverted indexing in Pubfig usingdifferent threshold T in Equation 8. When T is too large, we ignore theattribute signature so the performance will converge to SC. When T is toosmall, the performance will drop as it is sensitive to few noisy detectors.When using top 40 attributes ranked by ASC-W, the performance of attribute-embedded inverted indexing can be further improved.

several informative human attributes while still maintainingthe scalability of the system. We also find that using moreattributes does not guarantee better performance since someattributes will have negative effects on the performance (e.g.,Smiling, Mouth Closed). Therefore, we need to carefullyselect informative attributes identifying informative attributescomplementary to the low-level face feature which is anothercontribution for this work.

D. Experiments on attribute-embedded inverted indexing

Figure 7 shows the performance of attribute-embeddedinverted indexing in Pubfig using different threshold valuesT in Equation (8). When T is large, the performance willconverge to SC because it ignores the attribute signature. WhenT is small, the performance will be improved, but when Tis too small, the performance will drop dramatically. Thereare two possible reasons for this phenomenon. First, attributedetection error; when T is too small, the algorithm can nottolerate attribute detection error, so the performance will drop.Second, some attributes are not effective for identifying aperson, and these attributes will cause the performance dropwhen T is too small. To manifest the second point, we run thesame experiments on top 40 attributes ranked by the resultsof ASC-W. By removing some non-informative attributes, theperformance can be further improved from 16.3% to 17.6% inMAP (cf. Figure 7). This result also indicates the effectivenessof the attributes ranked by ASC-W.

E. Combining ASC-W and AEI

ASC-W and AEI can be combined to further improvethe retrieval performance. Table V shows the performanceof combination results. By combining two methods, we canachieve up to 43.55% and 42.39% relative improvement overthe original sparse coding. This is because ASC-W exploitsthe global structure of the feature space using only severalinformative attributes in the offline stage while AEI furtherconsiders locally to the designated query and just uses binarysignature of all informative attributes in the online stage. Notethat the performance gain is mainly contributed by ASC-W,but AEI is perfectly embedded into inverted index and is ableto reduce the retrieval time. According to evidence reported in

[35], such signatures can reduce the online retrieval time for 1million images from 0.62 seconds to 0.20 seconds. Besides theoverall MAP, precision on the top rank is also improved. Notethat our MAP is much lower than P@10 and P@20, this isbecause the content-based face retrieval generally suffers fromlow recall rate problem due to the semantic gap. AlthoughMAP is relatively low, the precision in top ranks is highenough for many applications; for instance, the system canbe combined with method proposed in [2] for automatic faceannotation.

F. Example results

Figure 8 (a) and (b) shows two query examples using SCand ASC-W respectively. The red boxes indicate the falsepositives, and the number below each image is its rank in theretrieval results and the number in parentheses represents therank predicted by SC. The improvement of ASC-W comparedwith SC is probably contributed by the attributes such as eyewears and hair colors. Note that even though some attributesare not always consistent with the same person (cf. Figure 8(b) top 1 in SC), we can still retrieve the images using ASC-W.Figure 8 (c) and (d) show two examples using the proposedmethods. Figure 8 (c) shows an example of occlusions. Usinghuman attributes like hair colors we can gather informationfrom not only face regions, therefore we can still achieve goodperformance under the occlusion. Figure 8 (d) shows a failurecase, because the quality of the query image is poor, we cannotcorrectly predict the human attributes and sparse codewords,therefore the performance is not improved using ASC-W.

G. Scalability

In this section, we discuss the scalability of our system intwo aspects, memory usage and online retrieval time.

1) Memory usage: In our current implementation, eachcodeword only needs 16 bits to store its image ID in theindex, and each image contains about 3,800 codewords onaverage. Total memory usage for inverted index with 13Kimages is about 94.2 MB. We also need to store 40 bitsattribute signature for each image (totally 0.1 MB for 13Kimages). Therefore, total memory usage is about 94.3 MBfor LFW dataset. For dataset with one million images, eachcodeword needs 20 bits to store its image ID; therefore, totalmemory usage for inverted index is about 9,060 MB. Forattribute signatures, they take about 4.8 MB. Total memoryusage is about 9,064.8 MB, which is a reasonable amount fora general computer server. Note that memory usage can befurther reduced by many compression techniques in informa-tion retrieval [14] (e.g., reducing around half of memory byadopting d-gap technique).

2) Online retrieval time: Our system is implemented usingC++ and operates on a 2.4GHz Intel Xeon server. For a singlequery, face detection and alignment take about 0.7 seconds,LBP feature extraction takes about 0.06 seconds, computingsparse representation takes about 0.35 seconds, and retrievingindex with 13K images takes about 0.03 seconds. For attributedetection, as shown in [36], detecting a single attribute can bedone within few milliseconds once the classifier is learned.For dataset with one million images, we refer to the results in

IEEE TRANSACTIONS ON MULTIMEDIA 10

TABLE VTHE RESULTS COMBINED WITH ASC-W AND AEI. BY COMBINING TWO METHODS TOGETHER, WE CAN ACHIEVE UP TO 43.55% RELATIVE

IMPROVEMENT IN LFW DATASET AND 42.39% IN PUBFIG DATASET.

LFW MAP Relative improv. Absolute improv. P@10 P@20SC 13.0% - - 46.8% 36.2%

ASC-W 18.3% 41.0% 5.3% 57.2% 45.3%AEI 14.3% 10.5% 1.3% 49.0% 37.5%

ASC-W+AEI 18.6% 43.6% 5.6% 57.3% 45.5%Pubfig MAP Relative improv. Absolute improv. P@10 P@20

SC 14.8% - - 49.0% 40.4%ASC-W 20.2% 36.9% 5.4% 56.4% 47.7%

AEI 17.6% 19.3% 2.8% 51.4% 42.8%ASC-W+AEI 21.0% 42.4% 6.2% 56.9% 48.3%

1

(SC)

(ASC-W)

2 3 4 5 6 7 8 9 10

1(2) 2(6) 3(10) 4(8) 5(80) 6(13) 7(7) 8(340) 9(113) 10(92)

(a)

1

(SC)

(ASC-W)

2 3 4 5 6 7 8 9 10

1(3) 2(6) 3(2) 4(1) 5(4) 6(7) 7(5) 8(15) 9(18) 10(31)

(b)

AP = 0.34

AP = 0.31

AP = 0.60

AP = 0.10

1(2) 2(120) 3(-) 4(1071) 5(55) 6(193) 7(-) 8(98) 9(16) 10(20)

(c)

AP = 0.11

1(8) 2(1) 3(-) 4(-) 5(-) 6(-) 7(-) 8(17) 9(-) 10(-)

(d)

AP = 0.06

Fig. 8. Example retrieval results using the proposed methods. The red boxes indicate the false positives, and the number below each image is its rank inthe retrieval results and the number in a parenthesis represents the rank by SC only. Attributes used in ASC-W including G:Male, H: Blond Hair, and P:NoEyewear (a) The improvement is probably due to that P: No Eyewear is one of the attributes included in ASC-W. (b) Similarly as probably including H:Blond Hair. Note that even if some attributes are not consistent within the same person, we might still be able to retrieve the correct images (e.g., Top 1 inSC and Top 4 in ASC-W) as using appearance codewords and other attributes. (c) Even the query image has some occlusions, we can still find certain resultsbecause the codewords and attributes can capture additional information from the face region. (d) The failure case is due to the poor quality for the queryimage such that we are unable to correctly detect human attributes and reliable codewords. (Best seen in color)

TABLE VICOMPARISON BETWEEN OUR INDEX STRUCTURE WITH THE ONE IN [35].

FOR A DATABASE CONTAINING ONE MILLION IMAGES, THE AVERAGELENGTH OF POSTING LIST IN [35] IS ABOUT 15,000 WHILE OUR SYSTEMIS ABOUT 13,571. BASED ON THE RESULTS REPORTED IN [35] AND THE

MEASURES BELOW, WE ARE ABLE TO RETRIEVE ONE MILLION FACEPHOTOS WITHIN ONE SECOND.

Baluja [35] Our system(a) Average # of codewords per image 3,000 3,800(b) Vocabulary size 200,000 280,000(c) Average length of posting list 15,000 13,571((a) × 1,000,000 images ÷ (b))

[35]. In their reports, retrieving index with one million imagestakes 0.2 seconds. Since we have similar index structure with[35] (cf. Table VI), we expect that retrieving index with onemillion face photos can be done in less than one second.

VI. CONCLUSIONS

We propose and combine two orthogonal methods to utilizeautomatically detected human attributes to significantly im-prove content-based face image retrieval (up to 43% relatively

in MAP). To the best of our knowledge, this is the first pro-posal of combining low-level features and automatically de-tected human attributes for content-based face image retrieval.Attribute-enhanced sparse coding exploits the global structureand uses several human attributes to construct semantic-awarecodewords in the offline stage. Attribute-embedded invertedindexing further considers the local attribute signature of thequery image and still ensures efficient retrieval in the onlinestage. The experimental results show that using the codewordsgenerated by the proposed coding scheme, we can reduce thequantization error and achieve salient gains in face retrievalon two public datasets; the proposed indexing scheme canbe easily integrated into inverted index, thus maintaining ascalable framework. During the experiments, we also discovercertain informative attributes for face retrieval across differentdatasets and these attributes are also promising for otherapplications (e.g., face verification). Current methods treat allattributes as equal. We will investigate methods to dynamically

IEEE TRANSACTIONS ON MULTIMEDIA 11

decide the importance of the attributes and further exploit thecontextual relationships between them.

REFERENCES

[1] Y.-H. Lei, Y.-Y. Chen, L. Iida, B.-C. Chen, H.-H. Su, and W. H. Hsu,“Photo search by face positions and facial attributes on touch devices,”ACM Multimedia, 2011.

[2] D. Wang, S. C. Hoi, Y. He, and J. Zhu, “Retrieval-based face annotationby weak label regularized local coordinate coding,” ACM Multimedia,2011.

[3] U. Park and A. K. Jain, “Face matching and retrieval using softbiometrics,” IEEE Transactions on Information Forensics and Security,2010.

[4] Z. Wu, Q. Ke, J. Sun, and H.-Y. Shum, “Scalable face image retrievalwith identity-based quantization and multi-reference re-ranking,” IEEEConference on Computer Vision and Pattern Recognition, 2010.

[5] B.-C. Chen, Y.-H. Kuo, Y.-Y. Chen, K.-Y. Chu, and W. Hsu, “Semi-supervised face image retrieval using sparse coding with identity con-straint,” ACM Multimedia, 2011.

[6] M. Douze and A. Ramisa and C. Schmid, “Combining Attributes andFisher Vectors for Efficient Image Retrieval,” IEEE Conference onComputer Vision and Pattern Recognition, 2011.

[7] N. Kumar, A. C. Berg, P. N. Belhumeur, and S. K. Nayar, “Describablevisual attributes for face verification and image search,” in IEEE Trans-actions on Pattern Analysis and Machine Intelligence (PAMI), SpecialIssue on Real-World Face Recognition, Oct 2011.

[8] W. Scheirer, N. Kumar, K. Ricanek, T. E. Boult, and P. N. Belhumeur,“Fusing with context: a bayesian approach to combining descriptiveattributes,” International Joint Conference on Biometrics, 2011.

[9] B. Siddiquie, R. S. Feris, and L. S. Davis, “Image ranking and retrievalbased on multi-attribute queries,” IEEE Conference on Computer Visionand Pattern Recognition, 2011.

[10] W. Scheirer and N. Kumar and P. Belhumeur and T. Boult, “Multi-Attribute Spaces: Calibration for Attribute Fusion and SimilaritySearch,” IEEE Conference on Computer Vision and Pattern Recognition,2012.

[11] G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller, “Labeledfaces in the wild: A database for studying face recognition in uncon-strained environments,” University of Massachusetts, Amherst, Tech.Rep. 07-49, October 2007.

[12] N. Kumar, A. C. Berg, P. N. Belhumeur, and S. K. Nayar, “Attributeand simile classifiers for face verification,” International Conference onComputer Vision, 2009.

[13] T. Ahonen, A. Hadid, and M. Pietikainen, “Face recognition with localbinary patterns,” European Conference on Computer Vision, 2004.

[14] J. Zobel and A. Moffat, “Inverted files for text search engines,” ACMComputing Surveys, 2006.

[15] A. Gionis, P. Indyk, and R. Motwani, “Similarity search in highdimensions via hashing,” VLDB, 1999.

[16] J. Sivic and A. Zisserman, “Video google: A text retrieval approachto object matching in videos,” International Conference on ComputerVision, 2003.

[17] D. Lowe, “Distinctive image features from scale-invariant keypoints,”International Journal of Computer Vision, 2003.

[18] O. Chum, J. Philbin, J. Sivic, M. Isard and A. Zisserman, “Total Recall:Automatic Query Expansion with a Generative Feature Model for ObjectRetrieval,” IEEE International Conference on Computer Vision, 2007.

[19] L. Wu, S. C. H. Hoi, and N. Yu, “Semantics-preserving bag-of-wordsmodels and applications,” Journal of IEEE Transactions on imageprocessing, 2010.

[20] Y.-H. Kuo, H.-T. Lin, W.-H. Cheng, Y.-H. Yang, and W. H. Hsu,“Unsupervised auxiliary visual words discovery for large-scale imageobject retrieval,” IEEE Conference on Computer Vision and PatternRecognition, 2011.

[21] J. Wang, S. Kumar, and S.-F. Chang, “Semi-supervised hashing forscalable image retrieval,” IEEE Conference on Computer Vision andPattern Recognition, 2010.

[22] D. Parikh and K. Grauman, “Relative Attributes,” IEEE InternationalConference on Computer Vision, 2011.

[23] Z. Cao, Q. Yin, J. Sun, and X. Tang, “Face recognition with learning-based descriptor,” IEEE Conference on Computer Vision and PatternRecognition, 2010.

[24] A. Torralba, K. P. Murphy, W. T. Freeman, and M. A. Rubin, “Context-based vision system for place and object recognition,” InternationalConference on Computer Vision, 2003.

[25] R. Raina, A. Battle, H. Lee, B. Packer, and A. Y. Ng, “Self-taughtlearning: Transfer learning from unlabeled data,” ICML, 2007.

[26] J. Yang, K. Yu, Y. Gong, and T. Huang, “Linear spatial pyramidmatching using sparse coding for image classification,” IEEE Conferenceon Computer Vision and Pattern Recognition, 2009.

[27] S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features:Spatial pyramid matching for recognizing natural scene categories,”IEEE Conference on Computer Vision and Pattern Recognition, 2006.

[28] J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma, “Robust facerecognition via sparse representation,” IEEE Transactions on PatternAnalysis and Machine Intelligence (PAMI), 2009.

[29] P. Viola and M. Jones, “Rapid object detection using a boosted cascadeof simple features,” IEEE Conference on Computer Vision and PatternRecognition, 2001.

[30] S. Milborrow and F. Nicolls, “Locating facial features with an extendedactive shape model,” European Conference on Computer Vision, 2008.

[31] J. Mairal, F. Bach, J. Ponce, and G. Sapiro, “Online dictionary learningfor sparse coding,” ICML, 2009.

[32] A. Coates and A. Y. Ng, “The importance of encoding versus trainingwith sparse coding and vector quantization,” ICML, 2011.

[33] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, “Least angleregression,” Annals of statistics, 2004.

[34] J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong, “Locality-constrained linear coding for image classification,” IEEE Conference onComputer Vision and Pattern Recognition, 2010.

[35] H. Jegou, M. Douze, and C. Schmid, “Hamming embedding andweak geometric consistency for large scale image search,” EuropeanConference on Computer Vision, 2008.

[36] Shumeet Baluja and H. Rowley, “boosting sex identification perfor-mance,” International Journal of Computer Vision, 2007.

Bor-Chun Chen received the M.S. degree from theDepartment of Computer Science and InformationEngineering, National Taiwan University, Taipei,Taiwan, in 2012. His research interests include mul-timedia analysis, large scale image retrieval, faceimage retrieval.

Yan-Ying Chen is currently working toward thePh.D. degree from the Department of Computer Sci-ence and Information Engineering, National TaiwanUniversity, Taipei, Taiwan. Her research interestsinclude face image analysis and multimedia datamining.

Yin-Hsi Kuo received the M.S. degree from theDepartment of Computer Science and InformationEngineering, National Taiwan University, Taipei,Taiwan, in 2010. She is currently working towardthe Ph.D. degree from the Graduate Institute ofNetworking and Multimedia, National Taiwan Uni-versity. Her research interests include multimediacontent analysis and image retrieval.

Winston H. Hsu (M’07-SM’12) received the Ph.D.degree in electrical engineering from ColumbiaUniversity, New York. He has been an AssociateProfessor in the Graduate Institute of Networkingand Multimedia, National Taiwan University, Taipei,Taiwan, since February 2007. Prior to this, he wasin the multimedia software industry for years. He isalso with the Department of Computer Science andInformation Engineering, National Taiwan Univer-sity, Taipei, Taiwan. His research interests includemultimedia content analysis, image/video indexing

and retrieval, machine learning, and mining over largescale databases. Dr.Hsu serves in the Editorial Board for the IEEE Multimedia Magazine.