2.5d face recognition using patch geodesic moments · most of the 2.5d face recognition studies...

14
2.5D face recognition using Patch Geodesic Moments Farshid Hajati a,b , Abolghasem A. Raie a , Yongsheng Gao b,c,n a Electrical Engineering Department, Amirkabir University of Technology, Tehran 15914, Iran b School of Engineering, Griffith University, QLD 4111, Australia c Queensland Research Lab, National ICT Australia, QLD 4207, Australia article info Article history: Received 30 September 2010 Received in revised form 9 August 2011 Accepted 24 August 2011 Available online 5 September 2011 Keywords: 2.5D face recognition Patch Geodesic Distance Patch Geodesic Moment Patch-based Geodesic distance Moment abstract In this paper, we propose a novel Patch Geodesic Distance (PGD) to transform the texture map of an object through its shape data for robust 2.5D object recognition. Local geodesic paths within patches and global geodesic paths for patches are combined in a coarse to fine hierarchical computation of PGD for each surface point to tackle the missing data problem in 2.5D images. Shape adjusted texture patches are encoded into local patterns for similarity measurement between two 2.5D images with different viewing angles and/or shape deformations. An extensive experimental investigation is conducted on 2.5 face images using the publicly available BU-3DFE and Bosphorus databases covering face recognition under expression and pose changes. The performance of the proposed method is compared with that of three benchmark approaches. The experimental results demonstrate that the proposed method provides a very encouraging new solution for 2.5D object recognition. Crown Copyright & 2011 Published by Elsevier Ltd. All rights reserved. 1. Introduction Face recognition has received significant amount of attention in recent years due to its numerous potential applications in the world. Much progress has been made toward recognizing 2D faces under controlled conditions, i.e. the model and probe images are taken in similar condition [1]. However, 2D face recognition encounters uncertainty problem in uncontrolled condition due to the inherent limitation of a 2D image for encoding the 3D nature of the human face/head. 3D information are being introduced in the research community to improve the recognition performance [2]. Compared to 3D face recognition (which collecting data from a 3601 scan), recognition from 2.5D data (which collecting range and texture data from a single viewing angle) is much appealing in real- world applications due to its very short image capturing time. The existing methods for 2.5D face recognition can be divided into two categories: range only approaches and range-texture approaches. The first category consists of methods which only use the range map of the 2.5D images for recognition. Cartoux et al. [3] used eigenfaces with automatically registered range map for face recognition from facial profiles. Gordon [4] explored the curvature features extracted from the range map. He represented the faces in terms of a vector of features and compared two faces based on their relationship in the feature space, achieving a recognition rate of above 80%. Tanaka et al. [5] considered the face recognition problem as a 3D shape recognition of free-form curved surfaces. They proposed a correla- tion-based face recognition approach by analyzing the maximum and minimum principal curvatures and their directions. Their experiments show that shape information from surface curvatures provides vital cues in discriminating and identifying complex sur- face structures as human faces. Chua et al. [6] extended the use of point signature to recognize frontal face scans with expression changes. The algorithm extracted the rigid parts of the face to create a model library for efficient indexing, which was evaluated on a small size range maps containing six human subjects, with four facial expressions. Pan et al. [7] utilized the partial directed Hausdorff distance to align and match two range maps for auto- matic 3D face verification. The experiments are carried out on a database with 30 individuals and a best EER of 3.24% is achieved. The second category consists of the algorithms, which use both range and texture maps to improve the recognition rate on 2.5D images. Chang et al. [8] showed that the face recognition perfor- mance on the texture map and the range map have no statistically significant difference in the rank-1 recognition rates (83.1% and 83.7% for the texture map and the range map, respectively). However, the combination of them can improve the recognition rate up to 92.8%, which is significantly greater than using either the texture map or the range map alone. Malassiotis and Strintzis [9] proposed a technique for face recognition using the range map to compensate pose variations in the corresponding texture map. The pose compensation on the input texture map was achieved by Contents lists available at SciVerse ScienceDirect journal homepage: www.elsevier.com/locate/pr Pattern Recognition 0031-3203/$ - see front matter Crown Copyright & 2011 Published by Elsevier Ltd. All rights reserved. doi:10.1016/j.patcog.2011.08.025 n Corresponding author at: School of Engineering, Griffith University, QLD 4111, Australia. Tel.: þ61 7 37353652; fax: þ61 7 37355198. E-mail address: yongsheng.gao@griffith.edu.au (Y. Gao). Pattern Recognition 45 (2012) 969–982

Upload: others

Post on 29-Jul-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

Pattern Recognition 45 (2012) 969–982

Contents lists available at SciVerse ScienceDirect

Pattern Recognition

0031-32

doi:10.1

n Corr

Australi

E-m

journal homepage: www.elsevier.com/locate/pr

2.5D face recognition using Patch Geodesic Moments

Farshid Hajati a,b, Abolghasem A. Raie a, Yongsheng Gao b,c,n

a Electrical Engineering Department, Amirkabir University of Technology, Tehran 15914, Iranb School of Engineering, Griffith University, QLD 4111, Australiac Queensland Research Lab, National ICT Australia, QLD 4207, Australia

a r t i c l e i n f o

Article history:

Received 30 September 2010

Received in revised form

9 August 2011

Accepted 24 August 2011Available online 5 September 2011

Keywords:

2.5D face recognition

Patch Geodesic Distance

Patch Geodesic Moment

Patch-based

Geodesic distance

Moment

03/$ - see front matter Crown Copyright & 2

016/j.patcog.2011.08.025

esponding author at: School of Engineering, G

a. Tel.: þ61 7 37353652; fax: þ61 7 37355198

ail address: [email protected] (Y

a b s t r a c t

In this paper, we propose a novel Patch Geodesic Distance (PGD) to transform the texture map of an

object through its shape data for robust 2.5D object recognition. Local geodesic paths within patches

and global geodesic paths for patches are combined in a coarse to fine hierarchical computation of PGD

for each surface point to tackle the missing data problem in 2.5D images. Shape adjusted texture

patches are encoded into local patterns for similarity measurement between two 2.5D images with

different viewing angles and/or shape deformations. An extensive experimental investigation is

conducted on 2.5 face images using the publicly available BU-3DFE and Bosphorus databases covering

face recognition under expression and pose changes. The performance of the proposed method is

compared with that of three benchmark approaches. The experimental results demonstrate that the

proposed method provides a very encouraging new solution for 2.5D object recognition.

Crown Copyright & 2011 Published by Elsevier Ltd. All rights reserved.

1. Introduction

Face recognition has received significant amount of attention inrecent years due to its numerous potential applications in theworld. Much progress has been made toward recognizing 2D facesunder controlled conditions, i.e. the model and probe images aretaken in similar condition [1]. However, 2D face recognitionencounters uncertainty problem in uncontrolled condition due tothe inherent limitation of a 2D image for encoding the 3D nature ofthe human face/head. 3D information are being introduced in theresearch community to improve the recognition performance [2].

Compared to 3D face recognition (which collecting data from a3601 scan), recognition from 2.5D data (which collecting range andtexture data from a single viewing angle) is much appealing in real-world applications due to its very short image capturing time. Theexisting methods for 2.5D face recognition can be divided into twocategories: range only approaches and range-texture approaches.The first category consists of methods which only use the range mapof the 2.5D images for recognition. Cartoux et al. [3] used eigenfaceswith automatically registered range map for face recognition fromfacial profiles. Gordon [4] explored the curvature features extractedfrom the range map. He represented the faces in terms of a vector offeatures and compared two faces based on their relationship in the

011 Published by Elsevier Ltd. All

riffith University, QLD 4111,

.

. Gao).

feature space, achieving a recognition rate of above 80%. Tanakaet al. [5] considered the face recognition problem as a 3D shaperecognition of free-form curved surfaces. They proposed a correla-tion-based face recognition approach by analyzing the maximumand minimum principal curvatures and their directions. Theirexperiments show that shape information from surface curvaturesprovides vital cues in discriminating and identifying complex sur-face structures as human faces. Chua et al. [6] extended the use ofpoint signature to recognize frontal face scans with expressionchanges. The algorithm extracted the rigid parts of the face to createa model library for efficient indexing, which was evaluated ona small size range maps containing six human subjects, withfour facial expressions. Pan et al. [7] utilized the partial directedHausdorff distance to align and match two range maps for auto-matic 3D face verification. The experiments are carried out on adatabase with 30 individuals and a best EER of 3.24% is achieved.

The second category consists of the algorithms, which use bothrange and texture maps to improve the recognition rate on 2.5Dimages. Chang et al. [8] showed that the face recognition perfor-mance on the texture map and the range map have no statisticallysignificant difference in the rank-1 recognition rates (83.1% and83.7% for the texture map and the range map, respectively).However, the combination of them can improve the recognitionrate up to 92.8%, which is significantly greater than using eitherthe texture map or the range map alone. Malassiotis and Strintzis[9] proposed a technique for face recognition using the range mapto compensate pose variations in the corresponding texture map.The pose compensation on the input texture map was achieved by

rights reserved.

Page 2: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982970

matching the input range map with a reference model using theICP algorithm [10]. The classification was performed using anEmbedded Hidden Markov Model (EHMM) [11] on the compen-sated texture map, which obtained encouraging system perfor-mance. Lu et al. [12] proposed a technique for matching between2.5D face images and 3D face models for face recognition underpose variations. For each subject a reference 3D face modelconstructed by merging several 2.5D face images from differentviewing angles. The recognition step consists of two components:a surface matching and an appearance matching. A modified ICPalgorithm is used to match the range map of the query 2.5D imageand the reference 3D models followed by an appearance matchingof texture maps.

Most of the 2.5D face recognition studies have focused on therange data, with a few using the texture data as supplementaryinformation to improve the recognition rate. In this paper, wepresent a novel Patch Geodesic Moments (PGM) and its applicationin 2.5D face recognition. The technique uses local texture momentscontrolled by the range information to tackle the problems of facerecognition under expression and pose variations when only oneexemplar 2.5D image per person is available. Unlike the otherresearches in 2.5D face recognition, our algorithm is applied on thetexture map to extract features with deformations controlled by therange map. A novel Patch Geodesic Distance is proposed to controlthe texture descriptor function, making it more robust to facevariations than the traditional geodesic distance and solves themissing data problem in 2.5D images.

The feasibility and effectiveness of the proposed method isinvestigated using the standard publically available BU-3DFEface database [13] and Bosphorus database [14], which containfaces under different expression changes and pose rotations.The performance of the system is compared with that of threestate-of-the-art benchmark approaches. Experimental resultsare very encouraging, which indicate that the proposedmethod provides a novel solution for face recognition using2.5D images.

The rest of this paper is organized as follows: Section 2 discussesbriefly the classical geodesic distance and its weakness, followed bya detailed description of our proposed Patch Geodesic Moments forobject representation and similarity measurement, and its applica-tion in 2.5D face recognition. In Section 3, the proposed method isextensively examined under various experimental conditions and

Dissimilarity Measurement

Decision

Gallery

Partitioning

2.5D Image Partitioned 2.5D Image

y

zx

Fig. 1. A flowchart of the

compared with three state-of-the-art approaches. Finally, the paperconcludes in Section 4.

2. Proposed Patch Geodesic Moments

In this paper, we propose a novel Patch Geodesic Moments (PGM)approach for 2.5D face recognition. It partitions the x–y plane (z axisgives depth information of the object) of a 2.5D image (whichconsists of a range map and a texture map) into equal-sized squarepatches in a non-overlapping manner on both range and texturemaps (see Fig. 1). A new Patch Geodesic Distance (PGD) is proposedto compute the shape of the object from the range map (see details inSection 2.2), which can handle the daunting missing data problem inthe classical geodesic distance calculation (see details in Section 2.1).The partitioned texture map is transformed via the PGD into atransformed texture map (see details in Section 2.3). This enablesthe texture invariance to viewing direction by aligning the texturemaps of the two 2.5D images to be matched. Then, the transformedtexture patches are coded and concatenated into a patch descriptorfor similarity measurement of 2.5D faces (see details in Sections2.4 and 2.5). A flowchart illustrating the proposed approach is givenin Fig. 1.

2.1. Geodesic distance

In graph theory, the distance between two vertices in a graph isthe number of edges in the shortest path connecting them. This isknown as the geodesic distance because it is the length of the graphgeodesic between those two vertices [15]. On a 3D object, the curvewith the minimum length between two points on the surface istheir geodesic path and its length is their geodesic distance. Thegeodesic distance have been used in a wide range of applicationsincluding fluid mechanics and computer vision [16]. Sethian [17]proposed a fast scheme for computing the geodesic distance on arectangular mesh. Kimmel and Sethian [18] extended the work bycomputing the geodesic distance on a triangulated mesh with thesame computational complexity. It provides an optimal-time algo-rithm for computing the geodesic distances on triangulated mani-folds. Hilaga et al. [19] explored the inter-geodesic distancesbetween points on a surface in order to match non-rigid surfaces.They defined a scalar function on the surface called distribution

Transformed Texture Map

Geodesic Texture

Transform

Patch Geodesic Distance

Computation

Patch Descriptor

Texture Map

Range Map

proposed approach.

Page 3: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

Fig. 2. Examples of the missing data problem on geodesic distance calculation.

(a) Change of the geodesic path due to self-occlusion, (b) the regions affected by

the occlusions resulting in incorrect geodesic distances, (c) change of the geodesic

path due to open mouth, and (d) the regions affected by the open mouth resulting

in incorrect geodesic distances.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982 971

function computed by integrating the geodesic distances from thegiven point to the rest of the points on the surface. The skeletalstructure of that scalar function was constructed and analyzed viaMultidimensional Reeb Graphs (MRG) [20]. The distance betweenMRGs was used as the dissimilarity measure. Elad and Kimmel [21]presented a method to construct a bending invariant signature forsurfaces based on geodesic distances between points. They mea-sured the inter-geodesic distances between uniformly distributedpoints on the surface. Then, a Multidimensional Scaling (MDS)technique was applied to extract coordinates in a finite dimensionalEuclidean space in which geodesic distances are replaced byEuclidean ones for matching.

Recently, the geodesic distance has been used as a feature in3D/2.5D face recognition algorithms [22–24]. Bronstein et al. [22]proposed an expression-invariant representation of faces calledCanonical Image Representation. This representation can modelthe deformations resulting from facial expressions. The canonicalimages are built by calculating the geodesic distances betweenpoints on the face surface. The experimental results demonstratedthat a smaller embedding error leads to better recognition.Mpiperis et al. [24] proposed a Geodesic Polar Representationof the face surface, which reduces the 3D face recognition tothe classification of expression-compensated 2D image. Theyachieved attractive recognition rates of 84.4% and 95% for twodifferent databases. These methods are appropriate for expres-sion-invariant representation, but fail on face surfaces in thepresence of missing data.

The computation of the geodesic distance on a surface involvessolving the Eikonal equation [25]:

9rTðQ Þ9¼ 1 ð1Þ

where T(Q) is the geodesic distance of the point Q(x, y, z) from areference point. If we consider an Eikonal equation in a 2D planewith a rectangular grid, the Eikonal equation can be approxi-mated by the following quadratic equation:

½maxðD�xij T,�Dþx

ij T,0Þ2þmaxðD�yij T,�Dþy

ij T,0Þ2�1=2 ¼ 1 ð2Þ

where Dgij is the differential operator at point (i, j) along the g

direction. The Fast Marching Method on Triangulated Domains(FMTD) [18] is the numerical approach to solve the Eikonal equationon surfaces. The main idea behind FMTD is to systematicallyadvance a front from the reference point in an upwind fashion toproduce the solution of T in every point of the triangulated grid.

The missing data on the surface can affect a geodesic pathresulting in a dramatic change of the geodesic distance of thepoint. This often occurs in 2.5D face images with in-depth facerotations and open mouth expressions. Examples of the problemare illustrated in Fig. 2. Fig. 2(a) and (c) shows the geodesic pathbetween a point on the face surface and the reference point (nosetip) and Fig. 2(b) and (d) depicts the shadowed regions in whichthe geodesic distance of the points from the reference point hasbeen changed.

2.2. Patch Geodesic Distance

To handle the above missing data problem in current geodesicdistance computation, we propose a Patch Geodesic Distance(PGD) computation approach and investigate its feasibility andeffectiveness for face recognition. The proposed PGD uses the localand global geodesic information on the surface to give a new wayof geodesic distance calculation for each surface point. It handlesmissing data by breaking a surface into a lower level (small scale),and then combine geodesic information at low and high levels. Itis a well-known fact that seeing things locally can help alleviateissues when addressed at a bigger scale. For example, Roweis andSaul [26] proposed the Local Linear Embedding (LLE) algorithm for

nonlinear dimensionality reduction, which represents nonlinearstructures in high dimensional data by exploiting the local symme-tries of linear reconstructions. Tenenbaum et al. [27] proposedan Isomap algorithm that uses local shortest paths in selectedneighborhoods of a nonlinear manifold for a global nonlineardimensionality reduction.

Given a 2.5D image, the Patch Geodesic Distance partitions thex–y plane of the 2.5D image (both range and texture maps) intoequal-sized square patches in a non-overlapping manner. For aN�N 2.5D image and a patch size of Wpatch, the number of patchesis (N/Wpatch)2. The PGD labels the patches with p and q indexes,which are integers ranging from 1 to N/Wpatch. A point (x, y, z) inthe 2.5D image is located in the (p, q)th patch, which can bedetermined as

p¼x

Wpatch

� �þ1 0rxoN ð3Þ

q¼y

Wpatch

� �þ1 0ryoN ð4Þ

where the symbol bUc denotes the floor function. Fig. 3 shows anexample of partitioning a 2.5D image into four patches.

To calculate the PGD for all surface points, we first compute aLocal Geodesic Distance for each point within its patch. Given apoint ðxpq,ypq,zpqÞ located in the (p,q)th patch, its local geodesicdistance Tl

pqðxpq,ypq,zpqÞ from a local reference point is calculated

Page 4: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

y

x

p=2, q=2

p=2, q=1p=1, q=1

y

z x

p=1, q=2

Fig. 3. An example of partitioning a 2.5D image into four patches. The partitioned range map is shown on the right side. For displaying purpose, we rescaled the range map

pixels to the range of [0,255]. (see Fig. 6 for the partitioned texture map).

Fig. 4. Computation of the angles of (a)ylðxpq ,ypqÞ and (b)yg

ðxLRPpq,yLRPpq

Þ in the range map.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982972

by solving the equation

9rTlpqðx

pq,ypq,zpqÞ9¼ 1, p,q¼ 1,. . .,N=Wpatch ð5Þ

where the superscript l stands for local geodesic distance. xpq, ypq,and zpq are the coordinates of the (p, q)th patch as

xpq ¼ x�Wpatchðp�1Þ ð6Þ

ypq ¼ y�Wpatchðq�1Þ ð7Þ

zpq ¼ f rpqðx

pq,ypqÞ ¼ f rðWpatchðp�1Þþxpq,Wpatchðq�1ÞþypqÞ ð8Þ

where f rðx,yÞ is the range map value of the given 2.5D image atpoint (x,y). Technically, any point in a patch can be used as thelocal reference point for that patch. Here, we define the centerpoint of a patch as its local reference point.

Secondly, we compute a Global Geodesic Distance for measur-ing the location of the patch in the partitioned 2.5D image. Theglobal geodesic distance is the geodesic distance of the patch’s localreference point from a global reference point. For the (p,q)th patch,the global geodesic distance TgðxLRPpq

,yLRPpq,zLRPpq

Þ, where the super-script g stands for global geodesic distance, is computed by solving

9rTgðxLRPpq,yLRPpq

,zLRPpqÞ9¼ 1 p,q¼ 1,. . .,N=Wpatch ð9Þ

where xLRPpq, yLRPpq, and zLRPpq are the coordinates of the localreference point of the (p,q)th patch.

Finally, the proposed PGD for any points on the surface iscalculated by combining the local and global geodesic distances.Let Q(x,y,z) be a point in the (p,q)th patch of a 2.5D image. ThePGD Patch T(Q) from the global reference point can be obtained by

PatchTðQ Þ ¼ ½ðTgðxLRPpq,yLRPpq

,zLRPpqÞcosyg

ðxLRPpq,yLRPpq

Þ

þTlðxpq,ypq,zpqÞcosylðxpq,ypqÞÞ

2

þðTgðxLRPpq,yLRPpq

,zLRPpqÞsinyg

ðxLRPpq,yLRPpq

Þ

þTlðxpq,ypq,zpqÞsinylðxpq,ypqÞÞ

2�1=2 ð10Þ

where ylðxpq,ypqÞ and yg

ðxLRPpq,yLRPpq

Þ are the angles of the linesconnecting between the point ðxpq,ypqÞ and the (p,q)th patch’slocal reference point, and between the (p,q)th patch’s localreference point and the global reference point, respectively. Theycan be obtained as

ylðxpq,ypqÞ ¼

0, if xpq ¼Wpatch

2 ,ypq ¼Wpatch

2

tan�1 ypq�Wpatch

2

xpq�Wpatch

2

� �, if xpq

ZWpatch

2

tan�1 ypq�Wpatch

2

xpq�Wpatch

2

� �þp, if xpqo Wpatch

2

8>>>>>><>>>>>>:

ð11Þ

ygðxLRPpq

,yLRPpqÞ ¼

0, if xLRPpq¼ xGRP ,yLRPpq

¼ yGRP

tan�1 yLRPpq�yGRP

xLRPpq�xGRP

� �, if xLRPpq

ZxGRP

tan�1 yLRPpq�yGRP

xLRPpq�xGRP

� �þp, if xLRPpq

oxGRP

8>>>><>>>>:

ð12Þ

where xGRP and yGRP are the coordinates of the global referencepoint. Fig. 4 illustrates how yl

ðxpq,ypqÞ and ygðxLRPpq

,yLRPpqÞ are

computed.Comparing to the traditional geodesic distance, the proposed

PGD can significantly reduce the areas where their geodesicdistances are affected by missing data on the surface. Fig. 5compares the effect of the missing data on the traditionalgeodesic distance and the proposed Patch Geodesic Distance. Itshould be noted that the traditional geodesic distance is a specialcase of our Patch Geodesic Distance when the patch size (Wpatch)is reduced to 1. Importantly, the proposed PGD is a generic 3Dshape measure applicable to any 3D surfaces.

As mentioned above, the main advantage of the proposedPatch Geodesic Distance is its robustness to missing data com-pared with the traditional geodesic distance. Here, we give an

Page 5: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

A

B

A

Bz

yx

S

S

2

O

Fig. 6. Computing the Patch Geodesic Distance on a hemispheric surface. (a) The 3D v

surface S, (c) the 3D view of the partitioned surface S with local reference points, and

indices.

Fig. 5. Comparing Patch Geodesic Distance with the traditional geodesic distance.

(a) A face surface with missing data, (b) regions with invalid traditional geodesic

distances, (c) the partitioned face surface in Patch Geodesic Distance, and

(d) regions with invalid Patch Geodesic Distances.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982 973

example to justify this advantage using Patch Geodesic Distanceformulation. Let S be a hemispheric surface having the radius gwith missing data, and A(g,g, g) and B(0.625g, 0.375g, 0.684g) aretwo points on S as shown in Fig. 6(a) and (b).

To compute the Patch Geodesic Distance between A and B, wefirst partition the surface S into g/2� g/2 square patches in thex–y plane using (3) and (4) and define the center point of eachpatch as the local reference point (see Fig. 6(c)). The local referencepoints and their (p,q) indices of the patches are shown in Fig. 6(d).

Given the point B(0.625g, 0.375g, 0.684g), the local geodesicdistance of B is computed by substituting p¼2, q¼1, andWpatch¼g/2 in (5)–(8) as

9rTl21ðx

21,y21,z21Þ9¼ 1 ð13Þ

x21 ¼ x�g2

ð14Þ

y21 ¼ y ð15Þ

z21 ¼ f r21ðx

21,y21Þ ¼ g2� x�g2

� �2

�y2

� �1=2

ð16Þ

where Tl21ðx

21,y21,z21Þ is the length of the shortest arc on thehemisphere between the local reference point of the (2,1)th patchand B as shown in Fig. 7(a). As can be seen, the missing data onthe hemispheric surface do not change the value of the localgeodesic distance for the point B.

By solving (13), the local geodesic distance becomes

Tl21ðx

21,y21,z21Þ ¼

Z o

0gdj¼ go ð17Þ

A

B

B

A

x

y

(2,4)

(2,1) (3,1)

(1,2)

(1,3)

(2,2)

(2,3)

(3,4)

(3,3)

(3,2)

(4,3)

(4,2)

2

O

iew of the surface S with missing data and points A and B, (b) the top view of the

(d) the top view of the partitioned surface S with local reference points and (p,q)

Page 6: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

B

O

Fig. 7. Computing the local geodesic distance and global geodesic distance on a surface. (a) Computing the local geodesic distance and (b) computing the global geodesic

distance.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982974

where o is the spatial angle between the line connecting thepoints B and the center point of the hemisphere and the lineconnecting the local reference point of the (2,1)th patch and thecenter point of the hemisphere.

The global geodesic distance of B, TgðxLRP21,yLRP21

,zLRP21Þ, is

computed by replacing p¼2 and q¼1 in (9) as

9rTgðxLRP21,yLRP21

,zLRP21Þ9¼ 1 ð18Þ

where xLRP21, yLRP21

, and zLRP21are the coordinates of the local

reference point of the (2,1)th patch.Eq. (18) is solved numerically on a triangulated mesh of the local

reference points as shown in Fig. 7(b). It can be seen that themissing data do not affect the triangulated mesh surface of the localreference points. Similar, the angles yl

ðx21,y21Þ and ygðxLRP21

,yLRP21Þ

are computed for B using (11) and (12) by the same substitution ofp¼2, q¼1, and Wpatch¼g/2, which are clear not affected by themissing data as well.

As the Patch Geodesic Distance (PGD) of the point B is computedby combining Tl

21ðx21,y21,z21Þ, TgðxLRP21

,yLRP21,zLRP21

Þ, ylðx21,y21Þ, and

ygðxLRP21

,yLRP21Þ using Eq. (10), the PGD is insensitive to the missing

data in this case. In general, the PGD with an appropriate patch sizeis always more robust (or insensitive) to missing data than theclassical geodesic distance (also see Fig. 5).

2.3. Geodesic texture transform

Having the Patch Geodesic Distance computed for all facepoints in Section 2.2, we can transform the texture map of a 2.5Dimage to a transformed texture map which is invariant to viewingdirection of the 2.5D object. Given a texture map f tðx,yÞ, itstransformed texture map f tðxtrans,ytransÞ is computed as

xtrans ¼ rðx,yÞcosjðx,yÞ ð19Þ

ytrans ¼ rðx,yÞsinjðx,yÞ ð20Þ

where r(x, y) is the Patch Geodesic Distance of the point Q(x, y,f r(x, y)). j(x, y) is the Geodesic Angle of the point (x, y) obtainedfrom the range map as

jðx,yÞ ¼

0,

if T

an

Tgð

tan�1 Tg ðxLRPpq ,yLRPpq ,zLRPpq ÞsinygðxLRPpq ,yLRPpq ÞþTlðxpq ,ypq ,zpqÞsinyl

ðxpq ,ypqÞ

Tg ðxLRPpq ,yLRPpq ,zLRPpq ÞcosygðxLRPpq ,yLRPpq ÞþTlðxpq ,ypq ,zpqÞcosyl

ðxpq ,ypqÞ

� �, if T

tan�1 Tg ðxLRPpq ,yLRPpq ,zLRPpq ÞsinygðxLRPpq ,yLRPpq ÞþTlðxpq ,ypq ,zpqÞsinyl

ðxpq ,ypqÞ

Tg ðxLRPpq ,yLRPpq ,zLRPpq ÞcosygðxLRPpq ,yLRPpq ÞþTlðxpq ,ypq ,zpqÞcosyl

ðxpq ,ypqÞ

� �þp, if T

8>>>>>>>>>>>><>>>>>>>>>>>>:

where Tlðxpq,ypq,zpqÞ, TgðxLRPpq,yLRPpq

,zLRPpqÞ, ylðxpq,ypqÞ, and yg

ðxLRPpq,

yLRPpqÞ are computed using (5), (9), (11), and (12), respectively.

With patching process, the (p, q)th patch in the partitionedtexture map, f t

pqðxpq,ypqÞ, can be represented as

f tpqðx

pq,ypqÞ ¼ f tðWpatchðp�1Þþxpq,Wpatchðq�1ÞþypqÞ ð22Þ

where xpq and ypq are the patch coordinates of the (p, q)th patch. p

and q are integers ranging from 1 to N/Wpatch computed from(3) and (4), respectively. Fig. 8 illustrates the partitioning of a2.5D image and its texture map into four patches.

By extending the above texture transform to a patch, thetransformed patch texture of the (p, q)th patch f t

pqðxpqtrans,y

pqtransÞ, is

calculated as

xpqtrans ¼ r Wpatchðp�1Þþxpq,Wpatchðq�1Þþypq

� �cosj Wpatchðp�1Þþxpq,Wpatchðq�1Þþypq

� �r Wpatchðp�1Þþ

Wpatch

2,Wpatchðq�1Þþ

Wpatch

2

� �

�cosj Wpatchðp�1ÞþWpatch

2,Wpatchðq�1Þþ

Wpatch

2

� �ð23Þ

and

ypqtrans ¼ rðWpatchðp�1Þþxpq,Wpatchðq�1ÞþypqÞsinjðWpatchðp�1Þ

þxpq,Wpatchðq�1ÞþypqÞ

�r Wpatchðp�1ÞþWpatch

2,Wpatchðq�1Þþ

Wpatch

2

� �

�sinj Wpatchðp�1ÞþWpatch

2,Wpatchðq�1Þþ

Wpatch

2

� �ð24Þ

Fig. 9 gives an example result of the patch geodesic texturetransform.

2.4. Patch descriptor

Moments have been utilized as a global feature in a number ofapplications [28,29]. Moment features are invariant under scaling,translation, rotation, and reflection. Moment invariants have beenproven to be successful in pattern recognition applications due to

gðxLRPpq,yLRPpq

,zLRPpqÞcosyg

ðxLRPpq,yLRPpq

ÞþTlðxpq,ypq,zpqÞcosylðxpq,ypqÞ ¼ 0

d

xLRPpq,yLRPpq

,zLRPpqÞsinyg

ðxLRPpq,yLRPpq

ÞþTlðxpq,ypq,zpqÞsinylðxpq,ypqÞ ¼ 0

gðxLRPpq,yLRPpq

,zLRPpqÞcosyg

ðxLRPpq,yLRPpq

ÞþTlðxpq,ypq,zpqÞcosylðxpq,ypqÞZ0

gðxLRPpq,yLRPpq

,zLRPpqÞcosyg

ðxLRPpq,yLRPpq

ÞþTlðxpq,ypq,zpqÞcosylðxpq,ypqÞo0

ð21Þ

Page 7: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

Patch Geodesic Texture

Transform

Fig. 9. An example of a partitioned texture map and its transformation.

y

x

p=2, q=2

p=2, q=1p=1, q=1

y

z x

p=1, q=2

Fig. 8. An example of partitioning a 2.5D image into four patches. The partitioned texture map is shown on the right side (see Fig. 3 for the partitioned range map).

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982 975

their sensitivity to pattern features. Hu [30] introduced the first setof algebraic moments invariants. Bamieh and De Figueiredo [31]derived Bamieh Moment Invariants (BMI), which had small featurevectors that were computationally efficient. Zernike Moments (ZM)[32] are orthogonal moments derived from the Zernike Polynomial[33]. Pseudo-Zernike Moments (PZM) is a modified version of ZMbased on another set of orthogonal polynomials. It is proved thatthe ZM and PZM have the best performance for image representa-tion among the common moments and the PZMs are more robustto noise than ZM [34]. ZMs and PZMs have been utilized asdescriptor in a range of face recognition applications in varyingexpression and lighting changing conditions, with partial occlu-sions, and recognition from one exemplar per person [35].

In this paper, we use PZMs as a patch descriptor in the proposedface recognition system. This patch descriptor is applied on thetransformed texture map created in the previous section to extractfeature vectors. Given a continuous intensity image f ðx,yÞ, PZM oforder n and repetition m can be defined as

PZMn,mðf ðx,yÞÞ ¼nþ1

p

ZZx2þy2 r1

Vn

n,mðx,yÞf ðx,yÞdxdy ð25Þ

where nZ0, 9m9rn, and the symbol n denotes the complexconjugate. The Pseudo Zernike polynomials Vn,m(x,y) are defined as

Vn,mðx,yÞ ¼ Rn,mðaÞejmy ð26Þ

where a¼ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffix2þy2

pis the length of the vector from the origin to the

pixel (x,y) and y¼ tan�1ðy=xÞ. The radial polynomials Rn,m(a) aredefined as

Rn,mðaÞ ¼Xn�9m9

s ¼ 0

ð�1Þsð2nþ1�sÞ!

s!ðn�9m9�sÞ!ðnþ9m9þ1�sÞ!an�s ð27Þ

By substituting (23) and (24) in (25), the PZM for the (p,q)thpatch in the transformed texture map f t xtrans,ytransð Þ is derived as

PZMpqn,mðf

tðxtrans,ytransÞÞ ¼nþ1

p

ZZxpq2

transþypq2trans r1

Vn

n,mðxpqtrans,y

pqtransÞ

� f tpqðx

pqtrans,y

pqtransÞdxpq

transdypqtrans ð28Þ

where p and q are integers ranging from 1 to N/Wpatch.To compute the PZM for each patch, the texture of the patch is

normalized into a unit circle. Because the texture of a patch has anirregular shape, we consider an optimum square bounding box tomap the patch into a unit circle for normalization. A zero paddingtechnique is used to pad the non-textural areas inside the boundingbox into zero. There are two possible approaches for this normal-ization. One is to map the bounding box in a way that the unit circleis inside the bounding box. In this way, the information of the pixelslocated outside the unit circle will be lost. The second approach,which we use in this paper, is to map the bounding box to bebounded inside the unit circle. This approach ensures that there isno pixel loss during the PZM computation. Fig. 10 illustrates anexample of the zero padding and the normalization processes.

Let the bounding box for the texture of a patch be a Wbox�Wbox

square. The (p,q)th patch of the transformed texture map ismapped into the unite circle as

xpqtrans,i ¼

iffiffiffi2p

Wboxi¼�

Wbox

2,�

Wbox

2þ1,. . .,

Wbox

2ð29Þ

ypqtrans,j ¼

jffiffiffi2p

Wboxj¼�

Wbox

2,�

Wbox

2þ1,. . .,

Wbox

2ð30Þ

where xpqtrans,i and ypq

trans,j are coordinates of the normalized (p,q)thpatch in the transformed texture map.

Eq. (28) is the continuous form of the PZM. Using Eqs. (29) and(30), we can derive a discrete form of PZM compatible for a digital

Page 8: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

1

1

x

y

1

1

x

y

Zero Padding

Normalization

Fig. 10. An example of the zero padding and the normalization processes.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982976

transformed texture map. Given the (p,q)th patch in a digitaltransformed texture map ft(xtrans,ytrans), the discrete PZM withorder n and repetition m is

PZMpqn,mðf

tðxtrans,ytransÞÞ

¼nþ1

plðWboxÞ

XWbox=2

i ¼ �Wbox=2

XWbox=2

j ¼ �Wbox=2

Vn

n,mðxpqtrans,i,y

pqtrans,jÞf

tpqðx

pqtrans,i,y

pqtrans,jÞ

ð31Þ

where the normalization factor l(Wbox) is the ratio between thenumber of pixels in the bounding box before and after mappinginto the unit circle. This normalization factor in our experimentsis W2

box=2. The magnitudes of PZMs with a minimum order nmin upto a maximum order nmax and repetition mZ0 are used as thefeature vector of the patch as shown below

PZMpqnmin ,nmax

¼ f9PZMpqu,vðf

tðxtrans,ytransÞÞ9���u¼ nmin,. . .,nmax and v¼ 0,. . .,ug

ð32Þ

where the symbol 9U9 denotes the absolute value of the vector.

2.5. Dissimilarity measurement

We use both the PZMs extracted from the transformed texturepatches and the location of each patch in the transformed texture

map to measure the dissimilarity between a query 2.5D objectand a model object. For this purpose, we compute two distancevectors. The first distance vector is Texture Distance Vector (TDV),which represents the textural difference between the two patchesin the query transformed texture map f and the model trans-formed texture map g:

TDVpq ¼ f½PZMpqnmin ,nmax

�f�½PZMpqnmin ,nmax

�gg ð33Þ

where PZMpqnmin ,nmax

is the PZM vector with the minimum order nmin

up to the maximum order nmax computed by (32).The second distance vector is Displacement Distance Vector

(DDV) which represents the spatial distance between the patchesin the query transformed texture map f and the model trans-formed texture map g:

DDVpq ¼TgðxLRPpq

,yLRPpq,zLRPpq

ÞcosygðxLRPpq

,yLRPpqÞ,

TgðxLRPpq,yLRPpq

,zLRPpqÞsinyg

ðxLRPpq,yLRPpq

Þ

" #f

8<:�

TgðxLRPpq,yLRPpq

,zLRPpqÞcosyg

ðxLRPpq,yLRPpq

Þ,

TgðxLRPpq,yLRPpq

,zLRPpqÞsinyg

ðxLRPpq,yLRPpq

Þ

" #g

9=; ð34Þ

where TgðxLRPpq,yLRPpq

,zLRPpqÞ and yg

ðxLRPpq,yLRPpq

Þ are computed from(9) and (12), respectively.

Because each patch after transformation has an irregularshape depending on its unique 3D shape, we consider this size

Page 9: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982 977

difference and make it into an adaptive weight to balance thecontribution of each patch in the similarity measurement. Theweight gpq for the (p,q)th patch is determined by the overlappedarea of the two patches to be matched as

gpq ¼SQ

pq \ SMpq

2ð35Þ

where SQpq and SM

pq are the areas of the (p,q)th patch aftertransforming the query and the model 2.5D images, respectively.Fig. 11 illustrates the computation of SQ

pq \ SMpq for the (p,q)th patch

in the query and model images.Using TDVpq, DDVpq, and gpq, we can compute Texture Distance

(Dtex) and Displacement Distance (Ddis) between the query and themodel 2.5D images as

Dtex ¼XN=Wpatch

p ¼ 1

XN=Wpatch

q ¼ 1

ðgpqU:TDVpq:Þ2

!1=2

ð36Þ

Ddis ¼XN=Wpatch

p ¼ 1

XN=Wpatch

q ¼ 1

ðgpqU:DDVpq:Þ2 !1=2

ð37Þ

1

1

x

y

1

1

y

Information Area Extraction

Mapping

Query Transformed Texture Map

Fig. 11. Computation of SQpq \ SM

pq for a patch in th

where the symbol :U: denotes the Euclidean norm of thevector.

Finally, the distance between the query 2.5D image (f) and themodel 2.5D image (g) is computed as

Dðf ,gÞ ¼ ðD2texþwdisUD2

disÞ1=2

ð38Þ

where wdis is a displacement weight to balance the contributionsof the textural and spatial disparities. For a given query 2.5Dimage, we compute its distance against all model 2.5D images inthe gallery. The model in the gallery with the minimum distanceD is considered as the correct match.

3. Experimental results

In order to evaluate the feasibility and effectiveness of theproposed approach, an extensive experimental investigationis conducted, covering 2.5D face recognition under facial expres-sion changes and pose rotations. The performance of the proposedmethod is compared with three state-of-the-art benchmarkapproaches: (1) the Geodesic Polar Representation method proposedby Mpiperis et al. [24] for expression-invariant face recognition,

x

1

1

x

y

Model Transformed Texture Map

e transformed query and model 2.5D images.

Page 10: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982978

(2) the Canonical Image Representation method proposed byBronstein for expression-invariant face recognition [22], and (3)the method of Malassiotis and Strintzis [9] for pose-invariant facerecognition.

The experiments were conducted on the BU-3DFE database[13], which contains 3D face models under different expressionchanges and the Bosphorus database [14] which contains 2.5Dscans of different poses. The Binghamton University 3D FacialExpression (BU-3DFE) database is tested in this paper to comparethe performance of our method with the benchmarks [22] and [24]under expression changes. The database contains 100 subjects (56females and 44 males) of 18 to 70 years old, with a variety ofethnic/racial ancestries. Each subject has seven expressions. Exceptthe neutral expression, the other six expressions (happiness,disgust, fear, angriness, surprise, and sadness) all include fourlevels of expression intensity. Therefore, there are 25 expressionmodels for each subject, resulting in a total of 2500 3D facialexpression models in the database. The Bosphorus database istested in this study to evaluate the performance of the proposedmethod under pose variations. It has over 5200 2.5D scans from105 subjects (61 females and 44 males) in different poses and facialexpressions. The data were acquired using Inspeck Mega CapturorII 3D [36]. The sensor resolution in x, y, and z (depth) dimensionsare 0.3 mm, 0.3 mm, and 0.4 mm respectively, and the colortexture maps have a resolution of 1600�1200 pixels.

Fig. 12. The effects of the patch size and the displacement weight on the

recognition rate.

Angriness Disgust Fear Hap

Level 1

Level 2

Level 3

Level 4

Fig. 13. An example set of faces for one

In all experiments, a preprocessing of face localization wasapplied. Original faces were normalized and cropped to the size of160�160 pixels in the way that the distance of the nose tip is80 pixels from the sides, 90 and 70 pixels from the top and thebottom, respectively.

3.1. Determination of parameters

In this section, we determine parameters involved in theproposed method, i.e. the minimum and the maximum ordersof the PZMs (nmin and nmax), the size of patches (Wpatch), and thedisplacement weight (wdis).

The PZMs are extracted from the transformed texture mapusing Eq. (32). In theory, the PZM with zero order represents theDC component (which can be proved by setting n¼0 in Eq. (31))and does not carry personal identity information. Hence, nmin isset as 1. nmax is taken to be 3 in our experiments [35].

The parameters Wpatch and wdis are determined experimentallyon a training dataset. The training dataset is created from theBU-3DFE database [13] as follows. For each subject, the neutral faceis selected as the gallery. The probe per person is randomly selectedfrom the remaining 24 faces with expressions of the subject. Thesystem performance with different values of Wpatch and wdis isdisplayed in Fig. 12. It can be seen that the recognition rate increaseswhen the patch size decreases (that is the image is partitioned intomore patches). The system reaches its best performance whenWpatch is 16 (that is the image is partitioned into 100 patches). Thenit drops when the patch size decreases further. Hence, the systemparameters are set as Wpatch¼16, wdis¼4.5, nmin¼1, and nmax¼3 forthe rest of the experiments.

3.2. Face recognition under facial expression changes

The robustness to facial expression variation is an importantissue in a face recognition system. In this experiment, we comparethe performance of the proposed method with that of the state-of-the-art methods [22,24] in this category on the BU-3DFE facedatabase [13]. The BU-3DFE face database has seven types ofexpression changes with four levels of expression intensities.Fig. 13 illustrates an example set of faces for one subject in thedatabase. In order to make a direct comparison with the resultsreported in [24], the same experimental strategy used in [24] isadopted in our experiments. The performance is measured interms of the rank-1 recognition rate and the Cumulative Match

piness Sadness Surprise Neutral

subject in the BU-3DFE database.

Page 11: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982 979

Characteristics (CMC) [37]. The neutral faces of all 100 people areused to form the single model per person gallery dataset, while therest images with expressions are used as probes. In total we have

Table 1Performance comparison under facial expression changes.

Expression The proposed

method

Geodesic Polar

Representation [24]a

Canonical Image

Representation [22]a

Angriness 93 N/A N/A

Disgust 79 N/A N/A

Fear 82 N/A N/A

Happiness 86 N/A N/A

Sadness 85 N/A N/A

Surprise 84 N/A N/A

Overall 84.8 80.3 77.2

a Results are from [24].

Fig. 14. Recognition rates obtained under facial expression changes.

+10° Rig+20° Right +30° Right

+45° Right

Right- Upwards

Right- Downwards

Fig. 15. An example set of rotated faces for on

six probe datasets, one for each expression with four differentexpression intensity levels, and on gallery dataset.

The rank-1 recognition rates of the proposed approach aretabulated in Table 1 together with the reported results of theGeodesic Polar Representation and Canonical Image Representation.It is encouraging to see from Table 1 that the proposed methodachieved an average accuracy of 84.8%, outperforming the state-of-the-art 3D expression-invariant techniques [24] and [22] by 4.5%and 7.6%, respectively. It is also observed that the recognition ratesfor different expressions range from 79% to 93%. The highest (93%)and the lowest (79%) accuracies are obtained from the angry anddisgusting expression faces, respectively. This shows that althoughthe geodesic-based methods can handle facial expression changeselegantly, the degree of face distortion caused by an expression stillhas certain impact on the recognition accuracy.

The CMC of the proposed method together with those of thetwo benchmark methods are given in Fig. 14. Note that therecognition rate of the proposed method is always higher thanthe benchmark methods.

3.3. Face recognition under pose rotations

The sensitivity to pose rotations is one of the most importantissues for the evaluation of face recognition systems. In thissection, we evaluate the robustness of the proposed methodagainst different face rotations on the Bosphorus database [14]which contains 2.5D face images in different poses. Fig. 15 gives anexample set of images for one subject in the database. In theexperiments, the frontal face is used as the single gallery of theperson, while other rotated faces are used as probes.

The recognition rates of the proposed approach for each rotationangles are tabulated in Table 2. For comparison purpose, therecognition rates of the method proposed by Masassiotis and Strintzis[9] are also given as the benchmark. The CMC curves of the proposedmethod and the benchmark method are shown in Figs. 16–26.

Frontal

Upwards (Slight)

Upwards (Strong)

Downwards (Slight)

Downwards (Strong)

-45° Left

ht

e subject in the Bosphorus face database.

Page 12: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

Table 2Performance comparison under pose rotations.

Rotation The proposedmethod

Malassiotis andStrintzis method [9]

þ101 Right 92.3 65.7

þ201 Right 88.6 52.4

þ301 Right 80 42.9

þ451 Right 39 9.5

�451 Left 38.1 21

Upwards (slight) 87.6 81.9

Upwards (strong) 79 69.5

Downwards (slight) 87.6 84.8

Downwards (strong) 69.5 68.6

Right-upwards 63.8 31.4

Right-downwards 34.3 21.9

Overall 69.1 50

Fig. 16. Recognition rate for faces under þ101 rotation.

Fig. 17. Recognition rate for faces under þ201 rotation.

Fig. 18. Recognition rate for faces under þ301 rotation.

Fig. 19. Recognition rate for faces under þ451 rotation.

Fig. 20. Recognition rate for faces under –451 rotation.

Fig. 21. Recognition rate for faces under slight upwards rotation.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982980

The proposed method achieves a 92.3% recognition rate for the þ101rotation to the right, while the benchmark has the recognition rate of65.7%. For the þ201 rotation, the recognition rates for the proposedmethod and the benchmark are 88.6% and 52.4%, respectively. Again,for the þ301 rotation, the recognition rate of the proposed methodstill outperforms the benchmark approach by 37.1%, showing itsconsistent superior performance in accuracy. Overall, the proposedmethod achieves an average accuracy of 69.1% across all rotationangles in the database, outperforming the benchmark approach (50%)by 19.1%, which demonstrates the effectiveness of the proposed patchgeodesic approach in recognizing rotated 2.5D faces. It is worthmentioning that above performance is obtained from a system thatits parameters are trained by a dataset different from the testing data.

Page 13: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

Fig. 22. Recognition rate for faces under strong upwards rotation.

Fig. 23. Recognition rate for faces under slight downwards rotation.

Fig. 24. Recognition rate for faces under strong downwards rotation.

Fig. 25. Recognition rate for faces under right-upwards rotation.

Fig. 26. Recognition rate for faces under right-downwards rotation.

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982 981

4. Conclusion

2.5D image recognition is appealing in real-world applicationsbecause it not only contains most of the 3D shape information ofthe object but also avoids the prohibitive long capturing time in3D scans by taking range information from a single viewing angleshot. However, self-occlusion of the object is inevitable in 2.5Ddata resulting in missing data on the object surface. This makescurrent 3D object representation and recognition techniques failto function properly. In this paper, we present a novel PatchGeodesic Distance (PGD) to address the missing data problem in2.5D shape description. By bringing the patch-based local patternidea in 2.5D object recognition, local geodesic paths withinpatches and global geodesic paths for patches are combined in a

coarse to fine hierarchical computation of PGD for each surfacepoint to tackle the missing data problem in 2.5D images. Shapeadjusted texture patches are encoded into local patterns, enablingsimilarity measurement between two 2.5D images with differentviewing angles and/or object shape deformations.

The algorithm has been evaluated on 2.5D face images andcompared with three state-of-the-art benchmarks. Two bench-marks have been used to evaluate the system performance underthe expression changes and the third benchmark has been used tocompare the performance of the proposed method under poserotations. It is a very encouraging finding that the Patch GeodesicDistance performs consistently superior to the standard geodesicdistance under the expression and pose changes. This researchreveals that the Patch Geodesic Moments provides a new solutionfor the 2.5D face recognition, which may also find its applicationin general 2.5D object representation and recognition.

Acknowledgments

This work is partially supported by Australian Research Council(ARC) under Discovery Grant DP0877929.

References

[1] A.F. Abate, M. Nappi, D. Riccio, G. Sabatino, 2D and 3D face recognition:a survey, Pattern Recognition Letters 28 (14) (2007) 1885–1906.

[2] V. Blanz, T. Vetter, Face recognition based on fitting a 3D morphable model,IEEE Transactions on Pattern Analysis and Machine Intelligence 25 (9) (2003)1063–1074.

Page 14: 2.5D face recognition using Patch Geodesic Moments · Most of the 2.5D face recognition studies have focused on the range data, with a few using the texture data as supplementary

F. Hajati et al. / Pattern Recognition 45 (2012) 969–982982

[3] J.Y. Cartoux, J.T. LaPreste, M. Richetin, Face authentication or recognition byprofile extraction from range images, in: Proceedings of the WorkshopInterpretation of 3D Scenes, 1989, pp. 194–199.

[4] G.G. Gordon, Face recognition based on depth and curvature features, in:Proceedings of the IEEE Conference on Computer Vision and Pattern Recogni-tion, 1992, pp. 108–110.

[5] H.T. Tanaka, M. Ikeda, H. Chiaki, Curvature-based face surface recognitionusing spherical correlation—principal directions for curved object recogni-tion, in: Proceedings of the Third International Conference on Face andGesture Recognition, 1998, pp. 372–377.

[6] C. Chua, F. Han, Y. Ho, 3D human face recognition using point signature, in:Proceedings of the Fourth International Conference on Automatic Face andGesture Recognition, 2000, pp. 233–238.

[7] G. Pan, Z. Wu, Y. Pan, Automatic 3D face verification from range data, in:Proceedings of the IEEE International Conference on Acoustics, Speech, andSignal Processing, 2003, pp. 193–196.

[8] K. Chang, K. Bowyer, P. Flynn, Multi-modal 2D and 3D biometrics for facerecognition, in: Proceedings of the IEEE International Workshop on Analysisand Modeling of Faces and Gestures, 2003, pp. 187–195.

[9] S. Malassiotis, M.G. Strintzis, Robust face recognition using 2D and 3D data:pose and illumination compensation, Pattern Recognition 38 (12) (2005)2537–2548.

[10] P. Besl, N. McKay, A method for registration of 3D shapes, IEEE Transactionson Pattern Analysis and Machine Intelligence 14 (2) (1992) 239–256.

[11] L.R. Rabiner, Tutorial on hidden Markov models and selected applications inspeech recognition, Proceedings of IEEE 77 (2) (1989) 267–296.

[12] X. Lu, A.K. Jain, D. Colbry, Matching 2.5D face scans to 3D models, IEEETransactions on Pattern Analysis and Machine Intelligence 28 (1) (2006)31–43.

[13] L. Yin, X. Wei, Y. Sun, J. Wang, M.J. Rosato, A 3D facial expression database forfacial behavior research, in: Proceedings of the Seventh International Con-ference on Automatic Face and Gesture Recognition (FGR), 2006, pp. 211–216.

[14] A. Savran, N. Alyuz, H. Dibeklioglu, O. C- eliktutan, B. Gokberk, B. Sankur,L. Akarun, Bosphorus database for 3D face analysis, in: Proceedings of theFirst COST 2101 Workshop on Biometrics and Identity Management (BIOID),2008, pp. 47–56.

[15] J. Bouttier, P. Di Francesco, E. Guitter, Geodesic distance in planar graphs,Nuclear Physics 663 (3) (2003) 535–567.

[16] J.A. Sethian, Level Set Methods and Fast Marching Methods Evolving Inter-faces in Computational Geometry, Fluid Mechanics, Computer Vision, andMaterials Science, Cambridge University Press, Cambridge, 1999.

[17] J.A. Sethian, A fast marching level set method for monotonically advancingfronts, Proceedings of the National Academy of Sciences 93 (4) (1996)1591–1595.

[18] R. Kimmel, J.A. Sethian, Computing geodesic paths on manifolds, AppliedMathematics 95 (15) (1998) 8431–8435.

[19] M. Hilaga, Y. Shinagawa, T. Kohmura, T.L. Kunii, Topology matching for fullyautomatic similarity estimation of 3D shapes, in: Proceedings of the 28th

Annual Conference on Computer Graphics and Interactive Techniques, 2001,pp. 203–212.

[20] G. Reeb, On the singular points of a completely integrable pfaff form or of anumerical function, Comptes Rendus de l’Academic Sciences. Paris 222 (1)(1946) 847–849.

[21] A. Elad, R. Kimmel, On bending invariant signatures for surfaces, IEEETransactions on Pattern Analysis and Machine Intelligence 25 (10) (2003)1285–1295.

[22] A.M. Bronstein, M.M. Bronstein, R. Kimmel, Expression-invariant representa-tions of faces, IEEE Transactions on Image Processing 16 (1) (2007) 188–197.

[23] A.M. Bronstein, M.M. Bronstein, R. Kimmel, A. Spira, Face recognition fromfacial surface metric, in: Proceedings of the Eighth European Conference onComputer Vision, 2004, pp. 225–237.

[24] I. Mpiperis, S. Malassiotis, M.G. Strintzis, 3-D face recognition with thegeodesic polar representation, IEEE Transactions on Information Forensicsand Security 2 (3) (2007) 537–547.

[25] Y.R. Tsai, L. Cheng, S. Osher, H. Zhao, Fast sweeping algorithms for a class ofHamilton–Jacobi equations, SIAM Journal on Numerical Analysis 41 (2)(2003) 673–694.

[26] S.T. Roweis, L.K. Saul, Nonlinear dimensionality reduction by locally linearembedding, Science 290 (1) (2000) 2323–2326.

[27] J.B. Tenenbaum, V.D. Silva, J.C. Langford, A global geometric framework fornonlinear dimensionality reduction, Science 290 (1) (2000) 2319–2323.

[28] C.H. Teh, R.T. Chin, On image analysis by the methods of moments, IEEETransactions on Pattern Analysis and Machine Intelligence 10 (4) (1988)496–513.

[29] A. Nabatchian, E. Abdel-Raheem, M. Ahmadi, Human face recognition usingdifferent moment invariants: a comparative study, in: Proceedings of theCongress on Image and Signal Processing, 2008, pp. 661–666.

[30] M. Hu, Visual pattern recognition by moment invariants, IRE Transactions onInformation Theory IT-8 (1) (1962) 179–187.

[31] R. Bamieh, R. De Figueiredo, A general moment-invariants/attributed-graphmethod for three dimensional object recognition from a single image, IEEEJournal of Robotics and Automation RA-2 (1) (1986) 31–41.

[32] M.R. Teague, Image analysis via the general theory of moments, Journal of theOptical Society of America 70 (1) (1980) 920–930.

[33] F. Zernike, Beugungstheorie des Schneidenverfahrens und seiner verbesser-ten Form, der Phasenkontrastmethode, Physica 1 (1) (1934) 689–704.

[34] P. Tzouveli, K. Ntalianis, S. Kollias, Human face watermarking based onZernike Moments, in: Proceedings of the Fifth IEEE International Symposiumon Signal Processing and Information Technology, 2005, pp. 399–404.

[35] H.R. Kanan, K. Faez, Y. Gao, Face recognition using adaptively weighted patchPZM array from a single exemplar image per person, Pattern Recognition41 (12) (2008) 3799–3812.

[36] 3D Mega Capturor. /http://www.inspeck.com/pages/3d-mega-capturor.htmlS.[37] S.A. Rizvi, P.J. Phillips, H. Moon, The FERET verification testing protocol for

face recognition algorithms, in: Proceedings of the International Conferenceon Face and Gesture Recognition, 1998.

Farshid Hajati received the B.Sc. degree in electronic engineering from Khaje Nasir University of Technology, Iran, in 2003 and M.Sc. and Ph.D. degrees in electronicengineering from Amirkabir University of Technology, Iran, in 2006 and 2011, respectively. From 2009 to 2010 he was a visiting scholar with the School of Engineering,Griffith University, Australia. From 2011, he has been a member of the Faculty of Electrical Engineering at Tafresh University in Iran. His research interests include digitalimage processing, pattern recognition, computer vision, neural networks, and face detection and recognition.

Abolghasem A. Raie received the B.Sc. degree in electronic engineering from Sharif University of Technology, Iran, in 1973 and M.Sc. and Ph.D. degrees in electronicengineering from University of Minnesota, USA, in 1979 and 1982, respectively. From 1983, he has been a member of the Faculty of Electrical Engineering at AmirkabirUniversity of Technology in Iran. His research interests are algorithm design, performance analysis, machine vision, sensor fusion, and mobile robots navigation.

Yongsheng Gao received the B.Sc. and M.Sc. degrees in electronic engineering from Zhejiang University, China, in 1985 and 1988, respectively, and the Ph.D. degree incomputer engineering from Nanyang Technological University, Singapore. Currently, he is an Associate Professor with the Griffith School of Engineering, Griffith University,Australia. He is also with National ICT Australia, Queensland Research Laboratory, leading the Biosecurity group. His research interests include face recognition, biometrics,biosecurity, image retrieval, computer vision, and pattern recognition.