semi-supervised face image retrieval using sparse coding...

4
Semi-Supervised Face Image Retrieval using Sparse Coding with Identity Constraint Bor-Chun Chen, Yin-Hsi Kuo, Yan-Ying Chen, Kuan-Yu Chu, and Winston Hsu National Taiwan University, Taipei, Taiwan { siriushpa, kuonini, yanying}@gmail.com, [email protected], [email protected] ABSTRACT We aim to develop a scalable face image retrieval system which can integrate with partial identity information to improve the retrieval result. To achieve this goal, we first apply sparse coding on local features extracted from face images combining with inverted indexing to construct an efficient and scalable face retrieval system. We then propose a novel coding scheme that refines the representation of the original sparse coding by using identity information. Using the proposed coding scheme, face images with large intra-class variances will still be quantized into similar visual words if they share the same identity. Experimental results show that our system can achieve salient retrieval results on LFW dataset (13K faces) and outperform linear search methods using well known face recognition feature descriptors. Categories and Subject Descriptors H.3.3 [Information Search and Retrieval]: Retrieval models General Terms Algorithms, Performance Keywords Face, retrieval, Sparse coding, Semi-supervised, Identity 1. INTRODUCTION The goal of face image retrieval is to find the ranking result from most to least similar face images in a face image database given a query face image. Such work has many applications in different areas. For instance, when it applies on personal multimedia, it can enable automatic face tagging and face image clustering; while in forensics, it can help with crime investigation. Face retrieval task is closely related to face recognition task, a problem has been investigated for decades. Pentland and Turk introduced the concept of eigenfaces [13] which has been widely used; Ahonen et al. proposed using local texture descriptors: local binary pattern (LBP) [10] for face recognition [2]. The descriptor proposed is efficient and has good performance. The difference between face recognition and face retrieval is that face recognition task requires completely labeled data in the training set, and it uses learning based approach to find classification result while neither training set nor learning process is needed in face retrieval task, and it provides a ranking result. Although descriptors used in face recognition can be directly applied on retrieval task, it is not trivial to apply it on an indexing system due to its high dimensionality. One might consider using traditional image retrieval system such as bag-of-words (BoW) model [12] constructed by local descriptors like SIFT features [7] for this problem. But when applying these methods on face images the performance is unsatisfactory. It is because face images have higher intra-class variance and these methods neglect the important spatial information in face images. Wu et al. proposed a face image retrieval framework using identity based quantization and multi-reference re-ranking [16] to solve this problem. Motivated by their work, we propose a framework using more general coding scheme with the help of partial identity information to form a semi-supervised face image retrieval system. Similar idea was discussed in [15] for semantic hashing. It is easy to see the work has its necessity because in most of application for face retrieval, we might have some side information about our database images, including identity information, gender information, social network, etc. It is important if we can properly use such (noisy and partially available) information in the retrieval system to improve the results. To the best of our knowledge, little work aims to deal with this problem. Recently, sparse coding has shown promising results on many different applications such as image denoising and classification. Raina et al. proposed a machine learning framework using unlabeled data to learn basis of sparse coding for classification task [11]. Yang et al. [17] further improved it by applying on SIFT descriptors along with spatial pyramid matching [6]. Gao et al. proposed a variant of sparse coding called laplacian sparse coding [4] by adding a locality constraint in the objective function. All of them show that sparse representations are well adapted to natural signals. Taking advantage of sparse coding, we introduce a Figure 1. (a) Because face images have high intra-class variance, patches from two bush images are quantized into different sparse representations ( ! " ,! $ ). (b) Augment face sparse coding with identity constraint. Even though the original sparse representations are different, the proposed coding scheme will propagate identity-related visual words between images of the same identity and construct more semantic sparse representations ( ! % " ,! % $ ). In this example, number of overlapped visual words increase from zero to three. *Area Chair: Kiyoharu Aizawa. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. MM’11, November 28 December 1, 2011, Scottsdale, Arizona, USA. Copyright 2011 ACM 978-1-4503-0616-4/11/11...$10.00.

Upload: others

Post on 26-Mar-2020

11 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Semi-Supervised Face Image Retrieval using Sparse Coding ...cmlab.csie.ntu.edu.tw/~sirius42/papers/mm11.pdf · Wu et al. proposed a face image retrieval framework using identity based

Semi-Supervised Face Image Retrieval using Sparse Coding with Identity Constraint

Bor-Chun Chen, Yin-Hsi Kuo, Yan-Ying Chen, Kuan-Yu Chu, and Winston Hsu National Taiwan University, Taipei, Taiwan

{ siriushpa, kuonini, yanying}@gmail.com, [email protected], [email protected]

ABSTRACT We aim to develop a scalable face image retrieval system which can integrate with partial identity information to improve the retrieval result. To achieve this goal, we first apply sparse coding on local features extracted from face images combining with inverted indexing to construct an efficient and scalable face retrieval system. We then propose a novel coding scheme that refines the representation of the original sparse coding by using identity information. Using the proposed coding scheme, face images with large intra-class variances will still be quantized into similar visual words if they share the same identity. Experimental results show that our system can achieve salient retrieval results on LFW dataset (13K faces) and outperform linear search methods using well known face recognition feature descriptors.

Categories and Subject Descriptors H.3.3 [Information Search and Retrieval]: Retrieval models

General Terms Algorithms, Performance

Keywords Face, retrieval, Sparse coding, Semi-supervised, Identity

1. INTRODUCTION The goal of face image retrieval is to find the ranking result from most to least similar face images in a face image database given a query face image. Such work has many applications in different areas. For instance, when it applies on personal multimedia, it can enable automatic face tagging and face image clustering; while in forensics, it can help with crime investigation. Face retrieval task is closely related to face recognition task, a problem has been investigated for decades. Pentland and Turk introduced the concept of eigenfaces [13] which has been widely used; Ahonen et al. proposed using local texture descriptors: local binary pattern (LBP) [10] for face recognition [2]. The descriptor proposed is efficient and has good performance. The difference between face recognition and face retrieval is that face recognition task requires completely labeled data in the training set, and it uses learning based approach to find classification result while neither training set nor learning process is needed in face retrieval task, and it provides a ranking result. Although descriptors used in face recognition can be directly applied on retrieval task, it is not trivial to apply it on an indexing system due to its high dimensionality. One might consider using traditional image retrieval system such as bag-of-words (BoW) model [12]

constructed by local descriptors like SIFT features [7] for this problem. But when applying these methods on face images the performance is unsatisfactory. It is because face images have higher intra-class variance and these methods neglect the important spatial information in face images. Wu et al. proposed a face image retrieval framework using identity based quantization and multi-reference re-ranking [16] to solve this problem. Motivated by their work, we propose a framework using more general coding scheme with the help of partial identity information to form a semi-supervised face image retrieval system. Similar idea was discussed in [15] for semantic hashing. It is easy to see the work has its necessity because in most of application for face retrieval, we might have some side information about our database images, including identity information, gender information, social network, etc. It is important if we can properly use such (noisy and partially available) information in the retrieval system to improve the results. To the best of our knowledge, little work aims to deal with this problem. Recently, sparse coding has shown promising results on many different applications such as image denoising and classification. Raina et al. proposed a machine learning framework using unlabeled data to learn basis of sparse coding for classification task [11]. Yang et al. [17] further improved it by applying on SIFT descriptors along with spatial pyramid matching [6]. Gao et al. proposed a variant of sparse coding called laplacian sparse coding [4] by adding a locality constraint in the objective function. All of them show that sparse representations are well adapted to natural signals. Taking advantage of sparse coding, we introduce a

Figure 1. (a) Because face images have high intra-class variance, patches from two bush images are quantized into different sparse representations ( !", !$ ). (b) Augment face

sparse coding with identity constraint. Even though the original sparse representations are different, the proposed coding scheme will propagate identity-related visual words between images of the same identity and construct more semantic sparse representations ( !%", !%$ ). In this example,

number of overlapped visual words increase from zero to three.

*Area Chair: Kiyoharu Aizawa.

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee.

MM’11, November 28 December 1, 2011, Scottsdale, Arizona, USA.

Copyright 2011 ACM 978-1-4503-0616-4/11/11...$10.00.

Page 2: Semi-Supervised Face Image Retrieval using Sparse Coding ...cmlab.csie.ntu.edu.tw/~sirius42/papers/mm11.pdf · Wu et al. proposed a face image retrieval framework using identity based

face retrieval system using inverted index built from sparse representation of face images. Although using sparse coding combined with inverted indexing results in an efficient retrieval framework, it does not take advantage of using identity information. To address this point, we propose a novel coding scheme to refine the result of sparse coding based on identity information. One might think our method is similar to [4], but they are essentially different. Firstly, no template features and large laplacian matrix is needed in our method. Secondly, since we quantize each person separately, we can scale up to as many people as we want. Last but not least, we integrate (partially available) identity information to automatically derive semantic-related visual words. Figure 1 illustrates the proposed method. Using sparse coding with identity constraint, a query will be able to retrieve most images with the same identity as long as it is similar to one of them. To sum up, our contributions include identifying the problem of semi-supervised face image retrieval, introducing a general coding scheme for face retrieval problem, and further improving the performance by using identity information in the coding scheme. Extensive experiments show that the proposed method significantly outperforms linear search methods (cf. Table 1).

2. SYSTEM FRAMEWORK For every face image in database, we first employ a frontal face cascade detector [14] to find the location of the face, and then active shape model [9] is applied to locating five facial components, including two eyes, two mouth corners and nose tip. The face image is then aligned using the locations of the eyes. After alignment, we define 5x7 grids around each facial component to get a total of 175 grids using the same procedure in [16]. From each grid, we extract a 59 dimension uniform LBP feature descriptor. All 175 descriptors then are first quantized into visual words separately by using sparse coding (Nonzero entries in the sparse representation are considered as visual words). Images with identity information are further refined using coding scheme described in section 3 to embed the information. Inverted index is built using the final sparse representation. Note that each descriptor from different grid locations will be quantized separately, hence two visual words will match only if it is extracted from the same grid location. Through this way, we can encode important geometric information of face into visual words. Given a query image, we apply the same detection, alignment, and feature extraction methods described above. The extracted features are then quantized into visual words by sparse coding and used for retrieving inverted index. In the retrieval stage, we simply use histogram intersection to compute similarity score. The system overview is shown in Figure 2.

3. SPARSE CODING WITH IDENTITY CONSTRIANT

3.1 Sparse Coding (SC) Traditional BoW model use k-means clustering algorithm to learn dictionary for quantization. The algorithm solves the following optimization problem:

∑=

−n

i

iiVD

Dvx1

2

,min (1)

subject to &'()*+,- = 1, ‖+,‖1 = 1, +, ≽ 0, ∀6 where 7 = 8)1, … , ):; is a dictionary matrix with the size of 59×K, each column represents a centroid (totally K); < =8+1, … , +=; is the centroid indicator matrix, each vi indicates that the original feature xi belongs to a centroid in D. The constraint &'()*+,- = 1 means each feature can only be assigned to one centroid. This constraint is considered to be too strict because some features might be at boundary of two or more centroids, therefore many people suggest relaxing the constraint and putting L1 regularization term on vi instead. This then turn into another optimization problem known as sparse coding:

11

2

,min i

n

i

iiVD

vDvx λ+−∑=

(2)

subject to ‖+,‖ ≤ 1, ∀6 This optimization problem can be efficiently solved by an online optimization algorithm [8]. We adopt this method to learn dictionary D for later use. Because we quantize feature descriptors from each grid separately, we have to learn different dictionary from them. Consequently, there are total 175 dictionaries learned. Once the dictionary is learned, we can fix D in the above formulation and minimize the objective function along with vi separately to find the sparse representation of each feature. When D is fixed, the optimization problem turns into a least square problem with L1 regularization. Since the L1 regularization term makes the objective function non-differentiable when vi

contains zero elements, we cannot solve it by standard unconstraint optimization method. This challenge has led to many literatures using different approaches to solve this problem. Here we use LARS algorithm [3] to solve this problem. Because of the L1 regularization term, each sparse representation vi

will only contain several nonzero elements out of K dimension. These nonzero entries are then considered as the visual words from the descriptor xi. Since there are 175 grid locations, total dictionary size is 175×K.

Figure 2. The proposed system framework. 59 dimension uniform local binary pattern feature descriptors are extracted from 175 grids in an aligned face image. These features then are quantized into visual words using sparse coding. Images with identity information are further refined via the proposed method. Finally, inverted index is built using the final sparse representation.

Page 3: Semi-Supervised Face Image Retrieval using Sparse Coding ...cmlab.csie.ntu.edu.tw/~sirius42/papers/mm11.pdf · Wu et al. proposed a face image retrieval framework using identity based

3.2 Sparse Representation Refinement using Identity Constraint (SC+I) Sparse representations computed from section inverted indexing, but it does not embed the identity Hence, the retrieval results suffer from low recall rate due to intra-class variance. Therefore, we introduce a method to reduce the intra-class variance by using identity informationthe sparse representation extracted from the images of identity should be alike. To achieve this, we propose identity constraint into the original sparse coding refine each sparse representation vi separately:

22

ˆ1ˆ)1(ˆmin

F

T

i

p

iiv

vVvDxi

γββ +−−+−

where Vp = [vp1,…,vpm] are m sparse representationfrom section 3.1 which share the same identity withparameter for adjusting weight between identity information and visual feature, and is another parameter for adjusting the sparsity of the result. Since there are two different parametereach term in the objective function is in different scaleto set the parameters. Hence we propose a simply normalize the objective, the results are as follow:

2

2

2

2

ˆ

1ˆ)1(

ˆmin

i

F

T

i

P

i

ii

v vm

vV

x

vDx

i

γββ +−

−+−

Although the objective function is non-differentiableobjective function without L1 regularizationdifferentiable and convex. Therefore we can still solve this problem by using generalized algorithms for L1 optimization problem. Specifically, we use a projected gradient algorithm [1] proposed by Schmidt. After refinement, sparse representations from the same identity will propagate visual words to each other, and query will be able to retrieve most images from the same identity similar to at least one of them (cf. Figure 1 (b)).

4. EXPERIMENTS

4.1 Experimental Setting

4.1.1 Data Set We use all the images from LFW datasetexperiments. There are total 13,233 face images among 5people in this dataset, while 1,680 people with two or more images and 12 people with 50 or more images. We take ten images each from these 12 people, total 120 images,set. Figure 3 shows some example images from the dataset. Throughout the experiments we use mean average precision (MAP) and precision at one (p@1) as our measurement.

4.1.2 Face Retrieval Baselines We use two kinds of face descriptors with linear search as baselines to compare with our system1. The first onthe face recognition descriptor proposed in [2]face into 7 by 7 fixed grids, extract 59 dimension uniform LBP from each grid, and then concatenate the features from each grid

1 We also implemented the method in [16] and found that the

performance relative to linear search system is the same but MAP of our linear search system is lower than the number they reported. The possible reason might be that we use different descriptors and query set.

Refinement using

section 3.1 are ready for embed the identity information.

suffer from low recall rate due to high a method to reduce

using identity information. Intuitively, from the images of the same

should be alike. To achieve this, we propose to add an to the original sparse coding problem to

ivγ (3)

representations computed which share the same identity with vi, is a

parameter for adjusting weight between identity information and is another parameter for adjusting the

sparsity of the result. Since there are two different parameters and ent scales, it is hard

the parameters. Hence we propose a simply way to as follow:

1

i

i

v

v (4)

differentiable, the regularization term is twice

herefore we can still solve this for L1 constraint

rojected scaled sub-

from the same identity visual words to each other, and query will be able

to retrieve most images from the same identity as long as it is

We use all the images from LFW dataset [5] for our 233 face images among 5,749 680 people with two or more

images and 12 people with 50 or more images. We take ten , total 120 images, as our query

shows some example images from the dataset. mean average precision

as our performance

with linear search as irst one is based on [2]. We divided the

, extract 59 dimension uniform LBP the features from each grid

and found that the performance relative to linear search system is the same but MAP of our linear search system is lower than the number they reported. The possible reason might be that we use different

to form a global descriptor with 2,891 dimensionbaseline – fixed grids (BF) in the following second one is the simple concatenationextracted in our proposed method, it has 175 gdimension feature, total 10,325 dimensionbaseline – component based (BC) in the For fair comparison, in the experiments with percent identity information, after retrievsearch in baselines, we perform re-rankinginformation to retrieve images from top 10 only use these images as ranking result

4.2 Sparse Coding Retrieval P In sparse coding, we need to decide the size of dictionarysparsity parameter . In our experiments, we

dictionary size from 100 to 400 with

Results are shown in Figure 4. We find out that if

set (around 10-1 to 10-2) the performanceWhen is big (around 100), it penamuch that all the entries in sparse representationtherefore, the MAP drops to zero. Whenthe performance drops faster with smallerbecause the sparse representation are soword lost its discriminative power. When is chosen properly, the performance of sparse coding

exceeds baselines because the baselinefeatures suffer from the curse of dimensionalityFigure 4 the component based baseline fixed grids baseline (BF) , this is component based baseline have many overlaparea where contains more information, the weight in the fixed grid method, and proper weight increasingthe performance agrees with the finding in experiments, we use K=100, =0.1 as our default parameters.

4.3 Sparse Coding with Identity Constraint Retrieval Performance In sparse coding with identity constraint, we need to decide

the weight between visual features and identity information, and

Figure 3. Example images from LFW dataset. Images in the same column are of the same person.

Figure 4. Comparison between sparse coding with linear search baselines. Sparse coding outperforms baselines in different dictionary size K, this is because the baselines suffer from curse of dimensionality.

891 dimension, we call it following experiments. The

concatenation of local descriptors , it has 175 grids with 59 dimension, and we call it

in the following experiments. the experiments with one hundred after retrieving images using linear

ranking by using the identity top 10 ranked identities and

use these images as ranking results. (BF+R, BC+R)

Retrieval Performance In sparse coding, we need to decide the size of dictionary K and

. In our experiments, we have tried different

range from 10-4 to 10-1.

We find out that if is properly

performance affects little by the size K. penalizes the nonzero entry so

representation become zero, When is small (around 10-4),

faster with smaller dictionary size, this is are so dense that each visual

properly, the performance of sparse coding

because the baselines with high dimensional dimensionality. Note that in

baseline (BC) outperforms the because that grids from

many overlaps in the central face more information, this overlapping acts like

thod, and proper weight increasing the performance agrees with the finding in [2]. In the following

=0.1 as our default parameters.

with Identity Constraint

In sparse coding with identity constraint, we need to decide

the weight between visual features and identity information, and

. Example images from LFW dataset. Images in the are of the same person.

between sparse coding with linear

search baselines. Sparse coding outperforms baselines in , this is because the baselines suffer

Page 4: Semi-Supervised Face Image Retrieval using Sparse Coding ...cmlab.csie.ntu.edu.tw/~sirius42/papers/mm11.pdf · Wu et al. proposed a face image retrieval framework using identity based

sparsity parameter . In our experiments, we have tried

range from 0.1~0.9 with range from 0.1~0.5

performs best when is 0.1 and is 0.15 in the case of one

hundred percent images with identity information.

small indicating that we weight toward the identity information.This is probably because the identity informationexperiments is reliable. The results are shown in see that sparse coding with identity constraint relative improvement in MAP over the baselinesfeature descriptors (BC+R). It suggests that effectively utilizes the identity information for the retrieval taskwhile maintaining an efficient and scalable index structure. In the case of partial information, the performance decreasewhen the information is less, but as long as there is someinformation, using sparse coding with identity constraint always outperforms using sparse coding only. Figure performance of sparse coding with identity constraint under different percentage of identity information comparecoding without identity constraint. Note that the query images do not carry any identity information. Intuitively, adding the identity constraint number of visual words in sparse representationretrieval time. Figure 6 shows the relationship between average number of visual words and performance by fix

adjusting . We find out that under the same sparsity with

coding, sparse coding with identity constraint can achieve much higher performance. Table 1 summarizes the performance of the and baselines under the best setting. Note that the performance of proposed method outperforms the linear search method in both cases while using an efficient index structure.

5. CONCLUSIONS We propose a face retrieval method based on inverted index structure with sparse coding and achieve better performance than linear search based on well-known face recognitionWe then show how to extend the proposed method into semisupervised case. In the experiments, we show that using the identity information, we can achieve higher performance while maintain the same sparsity. In the future, we will test the scalability of our system and combine more context information into the proposed framework to further boost up the performance.

6. REFERENCES [1] http://www.cs.ubc.ca/~schmidtm/Software/L1General.html[2] T. Ahonen, A. Hadid, and M. Pietikainen,

with local binary patterns. In ECCV, 2004.[3] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani. Least

angle regression. In ANN STAT, 2004.

Figure 5. Performance of sparse coding with identity information under different percentages of images with identity information. It shows that our method can leverage (partial) identity information.

In our experiments, we have tried different

range from 0.1~0.5 and find out it

is 0.15 in the case of one

identity information. The best is

rd the identity information. information used in our

in Table 1. We can identity constraint achieve 70%

relative improvement in MAP over the baselines using the same It suggests that our method

for the retrieval task index structure.

the performance decreases the information is less, but as long as there is some identity

information, using sparse coding with identity constraint always Figure 5 shows the

se coding with identity constraint under compares with sparse

Note that the query images do

will increase the in sparse representation and increase the hows the relationship between average

by fixing =0.1 and

out that under the same sparsity with sparse

identity constraint can achieve much

the proposed method that the performance of

ear search method in both

We propose a face retrieval method based on inverted index with sparse coding and achieve better performance than

recognition descriptors. to extend the proposed method into semi-

supervised case. In the experiments, we show that using the identity information, we can achieve higher performance while

, we will test the ine more context information

to further boost up the performance.

http://www.cs.ubc.ca/~schmidtm/Software/L1General.html T. Ahonen, A. Hadid, and M. Pietikainen, Face recognition

ECCV, 2004. B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani. Least

[4] S. Gao, I. W.-H. Tsang, L.-T. Chia, Pare not lonely – laplacian sparse codiclassification. In CVPR, 2010

[5] G. B. Huang, M. Mattar, T. Berg, and E. LearnedLabeled faces in the wild: A database for studying face recognition in unconstrained environments. In ECCV, 2008.

[6] S. Lazebnik, C. Schmid, and J. Ponce. Beyonfeatures: Spatial pyramid matching for recognizing natural scene categories. In CVPR, 2006.

[7] D. Lowe. Distinctive image features from scalekeypoints. In IJCV, 2003.

[8] J. Mairal, F. Bach, J. Ponce, and G. Sapiro. Online dictionary learning for sparse coding. In ICML, 2009.

[9] S. Milborrow and F. Nicolls. Locating Facial Features with an Extended Active Shape Model. In ECCV, 2008.

[10] T. Ojala, M. Pietikainen, and T. Maenpaa. Multiresolution gray-scale and rotation invariant texture classificationlocal binary patterns. In IEEE T PATTERN ANAL

[11] R. Raina, A. Battle, H. Lee, B. Packer, and A. Y. Ng. Selftaught learning: Transfer learning from unlabeled data. InICML, 2007.

[12] J. Sivic and A. Zisserman. Video Google: A text retrievalapproach to object matching in videos. In ICCV, 2003.

[13] M. Turk and A. Pentland. Eigenfaces for recognition. COGNITIVE NEUROSCI, 1991.

[14] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In CVPR, 2001.

[15] J. Wang, S. Kumar, and S.-F. Chang. Semihashing for scalable image retrieval.

[16] Z. Wu, Q. Ke, J. Sun and H.-Y. ShumRetrieval with Identity-Based Quantization andReference Re-ranking. In CVPR, 2010.

[17] J. Yang, K. Yu, Y. Gong, and T. Huang. Linear spatial pyramid matching using sparse coding for image classification.CVPR, 2009.

Table 1. Performance of the proposed method. outperforms linear search no matter information. The left three columns information while the right are information. Note that we did not compare the result that using partial information because the baselinewith partial identity labels.

Method BF BC SC

MAP 0.10 0.12 0.16

P@1 0.61 0.71 0.80

. Performance of sparse coding with identity

s of images with It shows that our method can effectively

Figure 6. Relationship between sparsity and performance in sparse coding and sparse coding with identity information. Using identity constraint can achieve much higher performance while maintaining roughly the same sparsity.

Chia, P. Zhao. Local features sparse coding for image

G. B. Huang, M. Mattar, T. Berg, and E. Learned-Miller. Labeled faces in the wild: A database for studying face recognition in unconstrained environments. In ECCV, 2008. S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags of features: Spatial pyramid matching for recognizing natural

D. Lowe. Distinctive image features from scale-invariant

Mairal, F. Bach, J. Ponce, and G. Sapiro. Online dictionary ng for sparse coding. In ICML, 2009.

Locating Facial Features with . In ECCV, 2008.

T. Ojala, M. Pietikainen, and T. Maenpaa. Multiresolution scale and rotation invariant texture classification with

IEEE T PATTERN ANAL, 2002. R. Raina, A. Battle, H. Lee, B. Packer, and A. Y. Ng. Self-taught learning: Transfer learning from unlabeled data. In

J. Sivic and A. Zisserman. Video Google: A text retrieval approach to object matching in videos. In ICCV, 2003. M. Turk and A. Pentland. Eigenfaces for recognition. In J

P. Viola and M. Jones. Rapid object detection using a

cascade of simple features. In CVPR, 2001. F. Chang. Semi-supervised

for scalable image retrieval. In CVPR, 2010. Shum. Scalable Face Image

Based Quantization and Multi-. In CVPR, 2010.

J. Yang, K. Yu, Y. Gong, and T. Huang. Linear spatial pyra-mid matching using sparse coding for image classification. In

ormance of the proposed method. Sparse coding outperforms linear search no matter with or without identity

three columns are without any identity information while the right are with 100% identity

not compare the result that using partial information because the baselines cannot deal

BF+R BC+R SC+I

0.33 0.42 0.72

0.61 0.71 0.86

Relationship between sparsity and performance in

sparse coding and sparse coding with identity information. n achieve much higher

performance while maintaining roughly the same sparsity.