laplacian object: one-shot object …milanfar/publications/conf/...laplacian object: one-shot object...

5
LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVING PROJECTION Sujoy Kumar Biswas and Peyman Milanfar Electrical Engineering Department University of California, Santa Cruz 1156 High Street, Santa Cruz, CA, 95064 ABSTRACT One shot, generic object detection involves detecting a sin- gle query image in a target image. Relevant approaches have benefitted from features that typically model the local simi- larity patterns. Also important is the global matching of local features along the object detection process. In this paper, we consider such global information early in the feature extrac- tion stage by combining local geodesic structure (encoded by LARK descriptors) with a global context (i.e., graph struc- ture) of pairwise affinities among the local descriptors. The result is an embedding of the LARK descriptors (extracted from query image) into a discriminatory subspace (obtained using locality preserving projection [1]) that preserves the lo- cal intrinsic geometry of the query image patterns. Experi- ments on standard data sets demonstrate efficacy of our pro- posed approach. Index Termsobject detection, locality preserving pro- jection, manifold learning, principal component analysis 1. INTRODUCTION In this paper, we focus on a particular variety of fine-grained detection [2] where the objective is to take a single query im- age as input and locate the pattern of the query image in a (usually bigger) target image as illustrated in Fig. 1. Past re- search work [3, 4] has shown that exemplar based detection strategies can work with laudable success, sometimes as good as training based approaches in coping with specific pose vari- ations. Relevant approaches benefitted from descriptors that typically model the local similarity patterns. Shechtman et al. [4] proposed a pattern matching scheme based on self- similarity [5, 6, 7, 8, 9, 10], the premise of which is that lo- cal internal layout of self-similar pixels are shared by visu- ally similar images. In [3], Seo et al., modeled local geomet- ric layout with Locally Adaptive Regression Kernel (LARK) [11] descriptors, and computed similarity between query and target features at each pixel of the target image by sweeping the query window over the target. One common aspect that emerges from such competing approaches is the emphasis on Thanks to XYZ agency for funding. Fig. 1. An overview of the query based object detection methodology where a query pattern is detected in a visually similar part of the target image (on right). the role of i) local geometry in developing features, and ii) global context for robust pattern matching. For example, in [3], Seo et al. have shown that projecting LARK descrip- tors on principal components results in salient features con- tributing to improved performance. But we argue that while projecting descriptors on a discriminatory subspace as in [3], it is imperative to take into account which descriptor comes from what spatial location so as to preserve the intrinsic local geometry (something that PCA can’t do). To put local infor- mation in such global context, we employ a graph based di- mensionality reduction technique where a global graph struc- ture comprising pairwise affinities (among neighboring pix- els) embeds the local descriptors into a low dimensional man- ifold in order to preserve locality for superior pattern detec- tion. The overall strategy to detect a query image in a target im- age follows a three step process — extracting local descrip- tors, computing salient features from them, and detecting the features in the target image (see Fig. 1). To avoid ambigu- ous matching between sets of features, we propose a global graph structure that incorporates spatial information corre- sponding to each descriptor, and we build this graph with pairwise affinities between neighboring descriptors. Section 2 discusses graph based dimensionality reduction, Section 3 in- troduces the notion of geodesic affinity, Section 4 presents de- tection strategy, Section 5 includes experimental results, and Section 6 draws conclusions.

Upload: hatruc

Post on 19-Apr-2018

234 views

Category:

Documents


4 download

TRANSCRIPT

Page 1: LAPLACIAN OBJECT: ONE-SHOT OBJECT …milanfar/publications/conf/...LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVING PROJECTION Sujoy Kumar Biswas and Peyman Milanfar

LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVINGPROJECTION

Sujoy Kumar Biswas and Peyman Milanfar

Electrical Engineering DepartmentUniversity of California, Santa Cruz

1156 High Street, Santa Cruz, CA, 95064

ABSTRACT

One shot, generic object detection involves detecting a sin-gle query image in a target image. Relevant approaches havebenefitted from features that typically model the local simi-larity patterns. Also important is the global matching of localfeatures along the object detection process. In this paper, weconsider such global information early in the feature extrac-tion stage by combining local geodesic structure (encoded byLARK descriptors) with a global context (i.e., graph struc-ture) of pairwise affinities among the local descriptors. Theresult is an embedding of the LARK descriptors (extractedfrom query image) into a discriminatory subspace (obtainedusing locality preserving projection [1]) that preserves the lo-cal intrinsic geometry of the query image patterns. Experi-ments on standard data sets demonstrate efficacy of our pro-posed approach.

Index Terms— object detection, locality preserving pro-jection, manifold learning, principal component analysis

1. INTRODUCTION

In this paper, we focus on a particular variety of fine-graineddetection [2] where the objective is to take a single query im-age as input and locate the pattern of the query image in a(usually bigger) target image as illustrated in Fig. 1. Past re-search work [3, 4] has shown that exemplar based detectionstrategies can work with laudable success, sometimes as goodas training based approaches in coping with specific pose vari-ations. Relevant approaches benefitted from descriptors thattypically model the local similarity patterns. Shechtman etal. [4] proposed a pattern matching scheme based on self-similarity [5, 6, 7, 8, 9, 10], the premise of which is that lo-cal internal layout of self-similar pixels are shared by visu-ally similar images. In [3], Seo et al., modeled local geomet-ric layout with Locally Adaptive Regression Kernel (LARK)[11] descriptors, and computed similarity between query andtarget features at each pixel of the target image by sweepingthe query window over the target. One common aspect thatemerges from such competing approaches is the emphasis on

Thanks to XYZ agency for funding.

Fig. 1. An overview of the query based object detectionmethodology where a query pattern is detected in a visuallysimilar part of the target image (on right).

the role of i) local geometry in developing features, and ii)global context for robust pattern matching. For example, in[3], Seo et al. have shown that projecting LARK descrip-tors on principal components results in salient features con-tributing to improved performance. But we argue that whileprojecting descriptors on a discriminatory subspace as in [3],it is imperative to take into account which descriptor comesfrom what spatial location so as to preserve the intrinsic localgeometry (something that PCA can’t do). To put local infor-mation in such global context, we employ a graph based di-mensionality reduction technique where a global graph struc-ture comprising pairwise affinities (among neighboring pix-els) embeds the local descriptors into a low dimensional man-ifold in order to preserve locality for superior pattern detec-tion.

The overall strategy to detect a query image in a target im-age follows a three step process — extracting local descrip-tors, computing salient features from them, and detecting thefeatures in the target image (see Fig. 1). To avoid ambigu-ous matching between sets of features, we propose a globalgraph structure that incorporates spatial information corre-sponding to each descriptor, and we build this graph withpairwise affinities between neighboring descriptors. Section 2discusses graph based dimensionality reduction, Section 3 in-troduces the notion of geodesic affinity, Section 4 presents de-tection strategy, Section 5 includes experimental results, andSection 6 draws conclusions.

Page 2: LAPLACIAN OBJECT: ONE-SHOT OBJECT …milanfar/publications/conf/...LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVING PROJECTION Sujoy Kumar Biswas and Peyman Milanfar

Query PCA LPP& Target FQ(1) & FT (1) FQ(2) & FT (2) FQ(1) & FT (1) FQ(2) & FT (2)

(a) (b) (c) (d) (e)

Fig. 2. Salient features shown after dimensionality reduction of LARK descriptors: (a) query & target images, (b)-(c) salientquery (target) features FQ (FT ) learnt by projecting descriptors HQ (HT ) along two dominant principal components, (d)-(e)same LARK descriptors projected along two dominant eigenvectors of LPP (one can notice finer local details in these features)

2. LOCALITY PRESERVING GRAPH BASEDDIMENSIONALITY REDUCTION

Consider the parameterized image surface S(xi) = xi, z(xi),where xi denotes the 2D coordinate vector xi = [xi1xi2 ]

T,having intensity z(xi). We represent the descriptors com-puted at location xi as l-dimensional vector hi ∈ Rl. Thedescriptors are computed densely from the query Q as well astarget image T over a uniformly spaced grid of xi’s. We de-note the number of descriptors computed this way from queryQ as NQ and the same from target T as NT , respectively.The descriptor vectors hi are placed column wise to definethe descriptor matrix for query as HQ = [h1,h2, . . . ,hNQ ],and the same for target as HT = [h1,h2, . . . ,hNT

]. So wehave HQ ∈ Rl×NQ and HT ∈ Rl×NT .

To distill the redundancy in densely computed descriptors,Seo et al. [3] have used PCA as a dimensionality reductiontechnique on the query descriptors HQ to obtain salient fea-tures for object detection. In contrast, we learn a discrim-inatory subspace v from Q such that the query descriptorshi, when projected on v, respect the local geometric pattern.In other words, if hi and hj , extracted from Q, are closelyspaced over the image manifold S then their projections vThi

and vThj on a subspace v should be close as well. This ob-jective is achieved by minimizing the following locality pre-serving projection (LPP) [1] cost function —

JLPP =1

2

∑ij

(vThi − vThj)2Kij . (1)

The cost function JLPP with our proposed affinity measureKij (described in Sec. 3) incurs heavy penalty if neighboringdescriptors hi and hj are mapped far apart. Upon simplifica-tion [1], expression (1) leads us to the final penalty as follows:

JLPP = vTHQLHTQv. (2)

The matrix D is a diagonal matrix defined as Dii =∑

j Kij ,and L = D−K is known as the graph Laplacian. The con-straint vTHQDHT

Qv = 1 is imposed on D to prevent highervalues of Dii from assigning greater “importance” to descrip-tor hi. Minimizing JLPP with respect to the aforementionedconstraint leads us to the generalized eigenvalue problem.The projection vector v that minimizes (2) with respect to theconstraint, is given by the minimum eigenvalue solution tothe following generalized eigenvalue problem:

HQLHTQv = λHQDHT

Qv. (3)

The desired set of eigenvectors which builds our low dimen-sional LPP subspace comprises the trailing d eigenvectorscomputed as a solution of (3). We collect the set of d eigen-vectors as columns of V = [v1,v2, . . . ,vd] ∈ Rl×d. Sincethe descriptors are densely computed they typically lie on alower dimensional manifold making d a small integer such as4 or 5. The descriptor matrices HQ and HT when projectedon V lead to salient features for query Q and target T repre-sented by FQ, FT respectively as follows:

FQ = VTHQ ∈ Rd×NQ ,FT = VTHT ∈ Rd×NT . (4)

In contrast to the LPP, the objective of PCA is to preserve theglobal geometric structure of the data by projecting the de-scriptors along the direction of maximal variation, and thuspays no attention to preserving finer details of local pattern.Fig. 2 illustrates the difference showing the salient query fea-tures FQ learnt with PCA versus LPP. The images in the fig-ure demonstrate that LPP is able to preserve greater details inthe salient features than PCA.

The image descriptors hi in our proposed object detec-tion framework are general and can be any (e.g., SIFT [12] orHOG [13]). However, we advocate the use of LARK descrip-tors [3, 14] because LARK is specifically designed to robustlyencode the spatial layout of the image manifold.

Page 3: LAPLACIAN OBJECT: ONE-SHOT OBJECT …milanfar/publications/conf/...LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVING PROJECTION Sujoy Kumar Biswas and Peyman Milanfar

3. LOCAL GEOMETRY REPRESENTATION WITHLOCAL GEODESIC COMPUTATION

In this Section, we present a geodesic view to construct thegraph affinity matrix K. The local geodesic distance betweenthe two neighboring (i.e., closely spaced) pixels xi and xj onthe image manifold S(xi) can be approximated [14] by thedifferential arc length dsij as follows:

ds2ij = dx2i1 + dx2

i2 + dz2,

= dx2i1 + dx2

i2 + (∂z

∂xi1

dxi1 +∂z

∂xi2

dxi2)2. (5)

The second step in Eq. (5) follows from chain rule of dif-ferentials. We can discretize (5) by means of the follow-ing two approximations: dxi1 ≈ ∆xi1j1 = xj1 − xi1 , anddxi2 ≈ ∆xi2j2 = xj2 − xi2 , i.e., ∆xi1j1 and ∆xi2j2 rep-resenting displacements along the two image-axes. The dis-cretized form of the differential arc length is given by the fol-lowing expression:

ds2ij ≈ ∆x2i1j1 +∆x2

i2j2 + . . .

(∆z

∆xi1j1

·∆xi1j1 +∆z

∆xi2j2

·∆xi2j2)2,

= ∆xTijCi∆xij +∆xT

ij∆xij ,

≈ ∆xTijCi∆xij , (6)

where ∆xij = [∆xi1j1 ∆xi2j2 ]T, and Ci is the local gradient

covariance matrix (also called as steering matrix in [3]) com-puted at xi. The second step follows from the matrix-vectorrepresentation of previous step, and the last step is written byconsidering the fact that ∆xT

ij∆xij is data independent andtrivial in a small local window [14].

However, straightforward computation of Ci based onraw image gradient at a single pixel may be too noisy. There-fore, we estimate Ci as an “average” gradient covariancematrix CΩi by collecting derivatives of the image signalz(xi) over a patch Ωi of pixels centered at pixel xi (see Fig.3(b)-3(c)) followed by an eigen-decomposition. Representingfirst derivatives as ∆i1z(m) and ∆i2z(m) along the two axesxi1 and xi2 , respectively, we write the final expression of CΩi

as follows:

CΩi =∑m∈Ωi

[∆i1z(m)

2∆i1z(m) ·∆i2z(m)

∆i1z(m) ·∆i2z(m) ∆i2z(m)2

],

= ν1u1u1T + ν2u2u2

T,

= (√ν1ν2 + ε)α(

√ν1 + τ

√ν2 + τ

u1u1T+

√ν1 + τ

√ν2 + τ

u2u2T), (7)

where, ν1 and ν2 are eigenvalues of CΩi corresponding toeigenvectors u1 and u2, respectively. Also in the derivationabove, ε, τ, α are regularization parameters to avoid numer-ical instabilities and kept constant throughout all the experi-ments in this paper at 10−7, 1 and 0.1 respectively.

Fig. 3. (a) Geodesic distance dsij between the points xi andxj on the signal manifold S(x) is shown. Since patch Ωi iscentered at xi as shown in (b) and (c), using definition CΩi

(7)to compute dsij would make dsij = dsij . To make dsij sym-metric, we make Ω common, as the circumscribing patch Ωij

between xi and xj in (d)-(e). One can enlarge the commonpatch by adding an annular width of ∆r pixels as required.

3.1. Building a Graph Laplacian with Geodesic Affinities

Next, we build a graph structure from query image using thequery descriptors as the graph nodes. We define the affini-ties between neighboring descriptor locations xi and xi asfollows,

Kij =

e−ds2ij

2σ2 when ∥xi − xj∥2 ≤ γ,

0 otherwise ,(8)

where σ is a smoothness parameter and γ a radius withinwhich we limit our affinity computation. Unfortunately, com-puting dsij in (8) using CΩj (7) would render dsij (and henceKij) non-symmetric. This is because the support Ωi of CΩi isdifferent from Ωj of CΩj (see Fig. 3(b)-3(c)), and upon usingEq. (7) to compute dsij we have dsij = dsji. However, Kis assumed symmetric in the derivation of LPP subspace (2).As a remedy to this technical predicament we make the sup-port of CΩj in (8) common for both xi and xj as shown by acircumscribing rectangle in the Fig. 3(d)-3(e). The commonsupport is denoted by Ωij and we denote the correspondinggradient covariance matrix as CΩij

. One can enlarge the rect-angular patch by adding an annular width of ∆r pixels (equalto 2 shown in Fig. 3). The final expression of the affinitybecomes the following:

Kij =

e−∆xT

ijCΩij∆xij

2σ2 when ∥xi − xj∥2 ≤ γ,

0 otherwise .(9)

A straightforward solution of taking average of Kij and Kji

to make K symmetric does not work in practice because suchaverage oversmooths the dominant structure pattern over theimage manifold. Next, we discuss the detection frameworkusing salient features FQ and FT .

Page 4: LAPLACIAN OBJECT: ONE-SHOT OBJECT …milanfar/publications/conf/...LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVING PROJECTION Sujoy Kumar Biswas and Peyman Milanfar

(a) (b) (c) (d) (e)

Fig. 4. (a) Query & target image, (b) resemblance map f(ρ) with PCA, (c) transformed correlation peaks f(ρ) with PCA, (d)resemblance map f(ρ) with LPP, (e) transformed correlation peaks f(ρ) with LPP. The sharp peaks of f(ρ) with less crowdingis visible in LPP results.

4. DETECTION FRAMEWORK

To detect a given query Q in a target image T , we sweepQ over T and in each target position xi (i.e., centered atpixel xi) we compute cosine similarity between FQ and FT

as follows: ρ(FQ,FTi) = trace(FT

QFTi

∥FQ∥F ∥FTi∥F

). Since ρ ∈[−1, 1], to suppress small correlation values further down andto boost the correlation peaks up we transform the ρ accordingto Lawley-Hotelling trace statistic [15, 16] as follows: f(ρ) =ρ2

1−ρ2 . Here, f(ρ) gives us a confidence or resemblance map[3], the same size as the target image, and we threshold themap with τ to achieve 99% confidence level from the empiri-cal probability density function (PDF) of f(ρ).

5. EXPERIMENTAL RESULTS

We have evaluated our methodology on two data sets, namely,UIUC car data set [17] and MIT-CMU face data set [18]. TheLARK kernel size is maintained at 9 × 9 giving rise to 81-dimensional descriptors. We reduce the dimensionality to asmall integer of roughly 5 by the use of LPP.

UIUC data contain gray-scale car images at same scale,and at multiple scales in the test set. MIT-CMU faces is also agray-scale image data set and we have evaluated our method-ology on the same subset of images as done by Seo et al. [3].The test set consists of 43 gray-scale images containing a to-tal of 149 frontal faces, occurring at various scales, and 20gray-scale images [3] having faces at unusually large (> 60)angular orientation. For multi-scale and multi-oriented detec-tion, we have scaled and rotated the query image and searchedfor the transformed query pattern in target.

The locality preserving projection turns out to be robustin detecting query pattern even in the presence of noise, lowresolution, pose variation and viewpoint changes. The resem-blance maps for the target image in Fig. 4(a) is shown inFig. 4(b) (for PCA) and Fig. 4(d) (for LPP). The correla-tion peaks resulting from LPP tend to be much sharper thanPCA. The apparent crowding of correlation peaks in PCA isabsent in LPP. Fig. 5 presents some results on UIUC car dataset. We have summarized the performances of our methodol-ogy in Table 1 in terms of equal-error detection rates (equalto recall rate when it becomes same as precision rate). The

Fig. 5. Detection results on UIUC Car data set with a singlequery image (top left)

Table 1. Equal Error Detection RatesApproaches Data Sets

UIUC Car MIT-Single Multi- CMUScale Scale Faces

Training Proposed 90.76 76.91 89.01Free Seo et al. [3] 85.39 74.51 86.58With Agarwal et al. [17] 77.01 44.00 -

Training Mutch et al. [19] 99.94 90.06 -Kapoor et al. [20] 94.00 93.50 -Lampert et al. [21] 98.50 98.60 -

Wu et al. [22] 97.60 - -

proposed methodology is also compared with training-basedapproaches in the table. We have consistently used a singlequery in both the data sets to arrive at these performances.

6. CONCLUSION

In this paper we have proposed an embedding technique forlocal descriptors in the context of one-shot object detection.We have further connected the idea of LPP with a geodesicnotion of affinity, thereby replacing the traditionally used heatkernel to define graph affinities. The proposed approach isgeneral enough to be integrated with any choice of descrip-tors. Finally, we have demonstrated the efficacy of this ap-proach with improved experimental results.

Page 5: LAPLACIAN OBJECT: ONE-SHOT OBJECT …milanfar/publications/conf/...LAPLACIAN OBJECT: ONE-SHOT OBJECT DETECTION BY LOCALITY PRESERVING PROJECTION Sujoy Kumar Biswas and Peyman Milanfar

7. REFERENCES

[1] H. Xiaofei, Y. Shuicheng, H. Yuxiao, P. Niyogi, andH. J. Zhang, “Face recognition using laplacianfaces,”Pattern Analysis and Machine Intelligence, IEEE Trans-actions on, vol. 27, no. 3, pp. 328–340, 2005.

[2] A. Berg, R. Farrell, A. Khosla, J. Krause, F. F. Li, J. Li,and S. Maji, “Fine-grained challenge 2013,” Available:https://sites.google.com/site/fgcomp2013/.

[3] H. J. Seo and P. Milanfar, “Training-free, generic ob-ject detection using locally adaptive regression kernels,”Pattern Analysis and Machine Intelligence, IEEE Trans-actions on, vol. 32, no. 9, pp. 1688–1704, 2010.

[4] E. Shechtman and M. Irani, “Matching local self-similarities across images and videos,” in ComputerVision and Pattern Recognition, 2007. CVPR’07. IEEEConference on. IEEE, 2007, pp. 1–8.

[5] C. F. Chen and Y. C. F. Wang, “Exploring self-similarities of bag-of-features for image classification,”in Proceedings of the 19th ACM international confer-ence on Multimedia. ACM, 2011, pp. 1421–1424.

[6] K. Chatfield, J. Philbin, and A. Zisserman, “Efficientretrieval of deformable shape classes using local self-similarities,” in Computer Vision Workshops (ICCVWorkshops), 2009 IEEE 12th International Conferenceon. IEEE, 2009, pp. 264–271.

[7] T. Deselaers and V. Ferrari, “Global and efficient self-similarity for object classification and detection,” inComputer Vision and Pattern Recognition (CVPR), 2010IEEE Conference on. IEEE, 2010, pp. 1633–1640.

[8] A. Vedaldi and A. Zisserman, “Self-similar sketch,”in Computer Vision–ECCV 2012, pp. 87–100. Springer,2012.

[9] O. Boiman, E. Shechtman, and M. Irani, “In de-fense of nearest-neighbor based image classification,” inComputer Vision and Pattern Recognition, 2008. CVPR2008. IEEE Conference on. IEEE, 2008, pp. 1–8.

[10] A. Vedaldi, V. Gulshan, M. Varma, and A. Zisserman,“Multiple kernels for object detection,” in ComputerVision, 2009 IEEE 12th International Conference on.IEEE, 2009, pp. 606–613.

[11] H. Takeda, S. Farsiu, and P. Milanfar, “Kernel regres-sion for image processing and reconstruction,” ImageProcessing, IEEE Transactions on, vol. 16, no. 2, pp.349–366, 2007.

[12] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal of ComputerVision, vol. 60, no. 2, pp. 91–110, 2004.

[13] N. Dalal and B. Triggs, “Histograms of oriented gradi-ents for human detection,” in Computer Vision and Pat-tern Recognition, 2005. CVPR 2005. IEEE ComputerSociety Conference on. IEEE, 2005, vol. 1, pp. 886–893.

[14] H. J. Seo and P. Milanfar, “Face verification using thelark representation,” Information Forensics and Secu-rity, IEEE Transactions on, vol. 6, no. 4, pp. 1275–1286,2011.

[15] M. M. Tatsuoka and P. R. Lohnes, Multivariate analysis:Techniques for educational and psychological research,Macmillan Publishing Co, Inc, 1988.

[16] T. Calinski, M. Krzysko, and W. Wołynski, “A compari-son of some tests for determining the number of nonzerocanonical correlations,” Communications in Statistics:Simulation and Computation, vol. 35, no. 3, pp. 727–749, 2006.

[17] S. Agarwal, A. Awan, and D. Roth, “Learning to detectobjects in images via a sparse, part-based representa-tion,” Pattern Analysis and Machine Intelligence, IEEETransactions on, vol. 26, no. 11, pp. 1475–1490, 2004.

[18] H. A. Rowley, S. Baluja, and T. Kanade, “Neuralnetwork-based face detection,” Pattern Analysis andMachine Intelligence, IEEE Transactions on, vol. 20,no. 1, pp. 23–38, 1998.

[19] J. Mutch and D. G. Lowe, “Multiclass object recogni-tion with sparse, localized features,” in Computer Visionand Pattern Recognition, 2006 IEEE Computer SocietyConference on. IEEE, 2006, vol. 1, pp. 11–18.

[20] A. Kapoor and J. Winn, “Located hidden random fields:Learning discriminative parts for object detection,” inComputer Vision–ECCV 2006, pp. 302–315. Springer,2006.

[21] C. H. Lampert, M. B. Blaschko, and T. Hofmann, “Be-yond sliding windows: Object localization by efficientsubwindow search,” in Computer Vision and PatternRecognition, 2008. CVPR 2008. IEEE Conference on.IEEE, 2008, pp. 1–8.

[22] B. Wu and R. Nevatia, “Simultaneous object detectionand segmentation by boosting local shape feature basedclassifier,” in Computer Vision and Pattern Recognition,2007. CVPR’07. IEEE Conference on. IEEE, 2007, pp.1–8.