exploring the use of iriscodes for presentation …rossarun/pubs/chenrossiriscodes...un-normalized...

9
Exploring the Use of IrisCodes for Presentation Attack Detection Cunjian Chen and Arun Ross Michigan State University, USA {cunjian, rossarun}@msu.edu Abstract Iris recognition systems are vulnerable to presentation attacks where an adversary uses artifacts such as cosmetic contact lenses, printed eye images, doll eyes, etc. to un- dermine the system. In this work, we investigate whether IrisCodes, that are commonly used for iris recognition, can be used for presentation attack detection. IrisCodes are bi- nary phasor features extracted from the geometrically nor- malized iris, where the annular iris region is transformed to a rectangular entity. We demonstrate that including pupil information in IrisCodes improves presentation attack detection performance. Further, we demonstrate that ex- tracting binary phasor information from the un-normalized iris image - referred to as OcularCode in this paper - can further boost the performance. Experiments involving both printed iris images and cosmetic contact lenses from a benchmark dataset suggest that the proposed methods based on binary phasor codes achieve promising results. 1. Introduction Iris recognition systems are vulnerable to different types of presentation attacks (PAs), where an adversary presents a fabricated artifact or altered trait to the iris sensor [16, 4]. The intent is often to obfuscate one’s own identity, cre- ate a synthetic identity, or to spoof another person’s iden- tity. Typically observed attacks include, but not lim- ited to, printed iris images, cosmetic contact lenses, and fake eyeballs [11, 10, 28]. In order to detect or deflect such attacks, numerous presentation attack detection (PAD) schemes have been developed in the literature, including sensor-based [3, 22] and image-based solutions [8, 19]. The latter set of solutions mainly extract morphological and tex- tural patterns of the iris image without leveraging additional hardware to, for example, capture optical or physiological properties of the eye [2]. Existing image-based iris PAD methods, especially those based on convolutional neural networks (CNNs) [7, 20, 18, 29, 9], have demonstrated significant performance on dif- ferent benchmark datasets. All of these approaches directly operate on the raw texture of iris images. In this work, we investigate the potential of using IrisCodes for presentation attack detection. IrisCodes are complex binary codes that are extracted from the normal- ized 1 iris region using Gabor wavelets [5]. They can be viewed as binarized Gabor phasor responses. They have been primarily used for iris recognition, where the goal is to determine the degree of similarity of dissimilarity between a pair of iris images. However, IrisCodes have also been used for other tasks also such as determining the gender of an individual [26], clustering iris images into different cat- egories [14] and designing iris cryptosystems for template security [15]. The objective of this work is to investigate the feasibility of using IrisCodes for presentation attack de- tection. In this regard, we study the following three questions: Can IrisCodes be used to detect presentation attacks? Can pupil information be successfully incorporated into IrisCodes in order to enhance PAD performance? Does the binary phasor code extracted from un- normalized iris (or the pre-normalized iris) result in better PAD accuracy than that of IrisCodes? The first question explores whether IrisCodes can be simul- taneously used for both iris recognition and presentation at- tack detection. The second and third questions aim to ex- amine the pros and cons of including the pupil and normal- ization step in developing iris PAD solutions. The rest of the paper is organized as follows. Section 2 presents a literature survey on recent developments in iris presentation attack detection using CNNs and the use of IrisCodes in different applications. Section 3 describes the proposed PAD framework used in this work. Section 4 dis- cusses the result of iris presentation attack detection on a benchmark dataset and provides some feature visualization results. Conclusions are drawn in Section 5. 1 Normalization refers to the conversion of the annular iris region into a rectangular entity. C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS), (Los Angeles, USA), October 2018.

Upload: others

Post on 05-Jul-2020

9 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

Exploring the Use of IrisCodes for Presentation Attack Detection

Cunjian Chen and Arun RossMichigan State University, USA{cunjian, rossarun}@msu.edu

Abstract

Iris recognition systems are vulnerable to presentationattacks where an adversary uses artifacts such as cosmeticcontact lenses, printed eye images, doll eyes, etc. to un-dermine the system. In this work, we investigate whetherIrisCodes, that are commonly used for iris recognition, canbe used for presentation attack detection. IrisCodes are bi-nary phasor features extracted from the geometrically nor-malized iris, where the annular iris region is transformedto a rectangular entity. We demonstrate that includingpupil information in IrisCodes improves presentation attackdetection performance. Further, we demonstrate that ex-tracting binary phasor information from the un-normalizediris image - referred to as OcularCode in this paper -can further boost the performance. Experiments involvingboth printed iris images and cosmetic contact lenses froma benchmark dataset suggest that the proposed methodsbased on binary phasor codes achieve promising results.

1. IntroductionIris recognition systems are vulnerable to different types

of presentation attacks (PAs), where an adversary presentsa fabricated artifact or altered trait to the iris sensor [16, 4].The intent is often to obfuscate one’s own identity, cre-ate a synthetic identity, or to spoof another person’s iden-tity. Typically observed attacks include, but not lim-ited to, printed iris images, cosmetic contact lenses, andfake eyeballs [11, 10, 28]. In order to detect or deflectsuch attacks, numerous presentation attack detection (PAD)schemes have been developed in the literature, includingsensor-based [3, 22] and image-based solutions [8, 19]. Thelatter set of solutions mainly extract morphological and tex-tural patterns of the iris image without leveraging additionalhardware to, for example, capture optical or physiologicalproperties of the eye [2].

Existing image-based iris PAD methods, especially thosebased on convolutional neural networks (CNNs) [7, 20, 18,29, 9], have demonstrated significant performance on dif-ferent benchmark datasets. All of these approaches directly

operate on the raw texture of iris images.In this work, we investigate the potential of using

IrisCodes for presentation attack detection. IrisCodes arecomplex binary codes that are extracted from the normal-ized1 iris region using Gabor wavelets [5]. They can beviewed as binarized Gabor phasor responses. They havebeen primarily used for iris recognition, where the goal is todetermine the degree of similarity of dissimilarity betweena pair of iris images. However, IrisCodes have also beenused for other tasks also such as determining the gender ofan individual [26], clustering iris images into different cat-egories [14] and designing iris cryptosystems for templatesecurity [15]. The objective of this work is to investigatethe feasibility of using IrisCodes for presentation attack de-tection.

In this regard, we study the following three questions:

• Can IrisCodes be used to detect presentation attacks?

• Can pupil information be successfully incorporatedinto IrisCodes in order to enhance PAD performance?

• Does the binary phasor code extracted from un-normalized iris (or the pre-normalized iris) result inbetter PAD accuracy than that of IrisCodes?

The first question explores whether IrisCodes can be simul-taneously used for both iris recognition and presentation at-tack detection. The second and third questions aim to ex-amine the pros and cons of including the pupil and normal-ization step in developing iris PAD solutions.

The rest of the paper is organized as follows. Section 2presents a literature survey on recent developments in irispresentation attack detection using CNNs and the use ofIrisCodes in different applications. Section 3 describes theproposed PAD framework used in this work. Section 4 dis-cusses the result of iris presentation attack detection on abenchmark dataset and provides some feature visualizationresults. Conclusions are drawn in Section 5.

1Normalization refers to the conversion of the annular iris region into arectangular entity.

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection,"Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 2: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

IrisDetection

IrisCropping

IrisEncoding

Ocular Image

OcularCode-ImaginaryUnnormalized Iris

IrisSegmentation

IrisNormalization

IrisEncoding

Normalized Iris without PupilIrisCode-Imaginary

OcularCode-Real

IrisCode-Real

+

+

IrisSegmentation

IrisNormalization

IrisEncoding +

Normalized Iris with Pupil Pupil-IrisCode-Imaginary Pupil-IrisCode-Real

IC-PAD

Pupil-IC-PAD

OC-PAD

Figure 1: Flowchart depicting the computation of the three distinct binary codes used in this work. These binary codes,obtained using Gabor filters, are extracted from (a) the normalized iris region without pupil (IrisCodes), (b) the normalizediris region with pupil (Pupil-IrisCodes), and (c) the unnormalized iris region (OcularCode). The PAD CNNs correspondingto these three sets of binary codes are termed as IC-PAD, Pupil-IC-PAD and OC-PAD, respectively. The ocular image on theleft is from [28].

2. Related WorkCNN-based Iris PAD: He et al. [7] proposed a Multi-

patch Convolutional Neural Network (MCNN) that oper-ated on iris patches extracted from the normalized iris im-age. Their reported experiments demonstrated much higheraccuracy than hand-crafted features. A similar approachwas proposed in [20] that performed contact lens classi-fication based on a CNN referred to as ContlensNet, thatwas also trained on image patches obtained from the nor-malized iris images. Furthermore, each sampled patch wasrotated through four different angles to augment the train-ing data. Unlike previous CNN-based techniques where thenormalized iris image was used, Pala and Bhanu [18] sam-pled iris patches directly from the un-normalized iris re-gion. Similarly, Hoffman et al. [9] sampled patches of size96×96 from the un-normalized iris region of size 300×300.Chen and Ross [2] proposed a multi-task PAD (MT-PAD)approach that could simultaneously perform iris detectionand iris presentation attack detection on a given image. Itobtained comparable performances on two different bench-mark datasets with the MCNN approach [7]. All the afore-mentioned methods directly operate on the raw pixel inten-sities and not on IrisCodes.

IrisCodes: IrisCodes have been typically for iris recog-nition and have demonstrated both efficiency and reliabil-ity [5, 25]. There has been a growing interest in usingthis binary code for other tasks. Mukherjee and Ross [14]utilized IrisCodes and k-means clustering to perform irisindexing. Galbally et al. [6] demonstrated the possibilityof reconstructing an iris image from the binary IrisCode.Adamovic et al. [1] presented a fuzzy commitment schemeto generate cryptographic keys based on IrisCodes. Tapia etal. [26] demonstrated that the same binary IrisCode used forrecognition could also be used for gender prediction. Theyreported an accuracy of 89% on the GFI dataset consisting

of 3,000 images by fusing the best bits of the of IrisCodescorresponding to the left and right eyes.

In this work, we explore whether IrisCodes are able tocapture details pertaining to presentation attacks. If this ispossible, then the purview of IrisCodes can extend beyondjust iris recognition.

3. Proposed Method3.1. Pre-processing

As can be seen in Figure 1, we consider three dif-ferent types of inputs that are processed by three dif-ferent CNNs: the IrisCode CNN (IC-PAD), the Pupil-IrisCode CNN (Pupil-IC-PAD) and the OcularCode CNN(OC-PAD). The pre-processing steps employed for extract-ing these inputs are different. While the generation of theIrisCode and Pupil-IrisCode involves iris segmentation andnormalization (the normalized iris is used), generation ofthe Pupil-IrisCode involves iris detection and cropping (theun-normalized iris is used).

3.1.1 Iris Segmentation and Normalization

In this approach, we first manually segment the iris by lo-calizing circular contours pertaining to the inner and outerboundaries of the iris region [17]. Then, we transform thesegmented iris from the cartesian coordinate domain (x, y)to a pseudo-polar domain (γ, θ) by adopting the rubber-sheet model (see Figure 2) proposed by Daugman [13]:

I(x(γ, θ), y(γ, θ))→ I(γ, θ). (1)

Here, I is the pixel intensity, and γ and θ are the unit inter-val and angle in the range [0, 1] and [0, 2π], respectively.x(r, θ) and y(r, θ) are the linear combinations of the in-ner (xi(θ), yi(θ)) and outer boundary points (xo(θ), yo(θ))

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 3: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

)y,(x ii )y,(x 00

Figure 2: Illustration of using Daugman’s rubber sheetmodel to transform the annular iris region to a rectangularentity. The image on the left depicts an un-normalized iris,while the one on the right depicts the normalized iris.

computed as,

x(γ, θ) = (1− γ)xi(θ) + γxo(θ), (2)

y(γ, θ) = (1− γ)yi(θ) + γyo(θ). (3)

Here, the subscripts i and o are used to denote inner andouter boundaries, respectively. The sampling frequency of(γ, θ) determines the final size of the normalized iris re-gion [26]. If the radius of inner boundary is reduced to zero,then the pupil region will also be included during the nor-malization (see Figure 2).

3.1.2 Iris Detection

Unlike iris segmentation that localizes circular contours per-taining to the inner and outer boundaries of the iris region,iris detection determines a rectangular boundary around theiris. The difference between iris detection and segmentationcan be observed in Figure 1. In this work, the iris detec-tor based on the Darknet framework [2] was used to obtaina rectangular iris region encompassing its outer boundary.Rather than using a small network for detection, we use alarge network with increased number of convolutional lay-ers to perform this task. Here, the Darknet-19 architecture,which has 19 convolutional layers, was used as the back-bone. All convolutional layers are sequentially followedby batch normalization (BN) and a Rectified Linear Unit(ReLU). The final detection layer uses a multi-task regres-sion loss function during training to simultaneously penal-ize the difference between the ground-truth and predictedbounding boxes, and maximize the probability of correctlyidentifying the object enclosed inside the bounding box asthe iris region. Examples of the iris detector’s outputs canbe seen in Figure 3. It is evident that this approach is ableto successfully operate in challenging scenarios involvingocclusion and unfavorable illumination where conventionaliris segmentation might fail.

3.2. Iris Encoding: Binary Code Generation

After the iris region in an input image has been local-ized, either a normalized (see Eqn. (1)) or an un-normalized

(a) Live Iris (b) Printed Iris

(c) Printed Iris (d) Contact Lens

Figure 3: Iris detection results on images from the LivDet-Iris-2017-Clarkson dataset illustrating live, printed and cos-metic contact lens samples. The bounding box, representedas a rectangle, is seen to encompass the iris.

iris image is obtained. This region is then converted to abinary phasor code. This is accomplished by convolvingthe normalized or un-normalized iris with complex Gaborfilters and transforming the ensuing phase response into abinary code [13, 5]. The binary code has both a real and animaginary component.

The iris encoding process can be mathematically de-scribed as [13]:

h{Re,Im} = sgn{Re,Im}

∫x

∫y

I(x, y)e−iω(θ0−y)

×e−(γ0−x)2/α2

e−(θ0−y)2/β2

xdxdy. (4)

Here, h{Re,Im} is a complex value bit, which is based on thesign of the real and imaginary components of the integral.I(x, y) is the normalized or un-normalized iris image and(γ0, θ0) represents the the centre frequency of the Gaborfilter. α and β are the width and length of the filter.

Note that when the binary code is computed from theunnormalized iris image, the encoded image will have thesame size as the originally cropped iris image since no nor-malization is involved. On the other hand, the size of theencoded image from the normalized iris image is deter-mined by the sampling frequency related to Eqn. (1), whichis 64×256 for IC-PAD and 128×256 for Pupil-IC-PAD. Ascan be seen in Figure 4, more structural information is pre-served in the binary code when a live iris sample is encoded,compared to a printed sample.

3.3. Iris PAD

Once the binary code is generated, it is used by a CNNto learn features for presentation attack detection. The

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 4: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

(a) Live Iris (b) Real Part ofLive Iris

(c) Imaginary Partof Live Iris

(d) Printed Iris (e) Real Part ofPrinted Iris

(f) Imaginary Partof Printed Iris

(g) Contact Lens (h) Real Part ofContact Lens

(i) Imaginary Partof Contact Lens

Figure 4: The OcularCode of un-normalized live, printedand contact lens iris samples from the LivDet-Iris-2017-Clarkson dataset. Both “Live” and “Printed” iris samplesare from the same subject.

CNN-based features have shown superior performancesthan hand-crafted features on different presentation attackdatasets [7]; hence the choice of using CNN to automat-ically learn the features from IrisCodes. Note that threedifferent CNNs are developed: IC-PAD, Pupil-IC-PAD andOC-PAD. All three CNNs are based on Darknet-19 [21],that was previously used for ImageNet classification. Thisnetwork architecture has 19 convolutional layers and 5 maxpooling layers (see Table 1). It uses successive 3 × 3 and1×1 convolutional filters. Each convolution block is a com-posite function of three different operations: convolution,Batch Normalization (BN), followed by a Rectified LinearUnit (ReLU). Each max pooling layer reduces the originaldimensional size by a factor of 2. The final convolutionallayer, after reducing by a factor of 32, is 14×14×2. Globalaverage pooling (GAP) layer is used to average the featuremaps (14 × 14) from the last convolutional layer, to inferthe probability of being “live” or “PA”. Then the resultingvector is fed into the Softmax layer to compute the classprobabilities. The Softmax layer is defined as:

yi =eW

Ti xi+bi∑

j ewT

j xi+bj, (5)

Table 1: Summary of parameters used by the proposed PADnetwork. The convolution block consists of convolution,batch normalization and activation layers. All three CNNsuse the same network architecture.

Type of Layer Filters Size/Stride Output DimensionsConv0 32 3× 3/1 448× 448× 32Max 2× 2/2 224× 224× 32

Conv1 64 3× 3/1 224× 224× 64Max 2× 2/2 112× 112× 64

Conv2 128 3× 3/1 112× 112× 128Conv3 64 1× 1/1 112× 112× 64Conv4 128 3× 3/1 112× 112× 128Max 2× 2/2 56× 56× 128

Conv5 256 3× 3/1 56× 56× 256Conv6 128 1× 1/1 56× 56× 128Conv7 256 3× 3/1 56× 56× 256Max 2× 2/2 28× 28× 256

Conv8 512 3× 3/1 28× 28× 512Conv9 256 1× 1/1 28× 28× 256Conv10 512 3× 3/1 28× 28× 512Conv11 256 1× 1/1 28× 28× 256Conv12 512 3× 3/1 28× 28× 512Max 2× 2/2 14× 14× 512

Conv13 1024 3× 3/1 14× 14× 1024Conv14 512 1× 1/1 14× 14× 512Conv15 1024 3× 3/1 14× 14× 1024Conv16 512 1× 1/1 14× 14× 512Conv17 1024 3× 3/1 14× 14× 1024Conv18 2 1× 1/1 14× 14× 2

GAP 2Softmax 2

where xi,Wj ,Wi are, respectively, the i-th training sample,and the j-th and i-th columns of weight matrix W . bi andbj are the corresponding biases. yi is the computed proba-bility value from the Softmax layer. The following L2 lossfunction is used to compute the classification cost:

E =1

2N

N∑i=1

||yi − yi||2, (6)

where, N is the number of samples in a mini-batch, andyi and yi are, respectively, the estimated and ground-truthprobability values for individual samples.

3.4. Implementation Details

Due to limited number of training samples available instandard iris presentation attack datasets, we adopt a trans-fer learning scheme to train the proposed PAD framework.Since the Darknet-19 architecture is employed as the back-bone network, we use the Darknet-19 weights trained fromImageNet to initialize the network training. We use mini-

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 5: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

batch gradient descent with a starting learning rate of 0.001,weight decay of 0.0005 and momentum of 0.9 to train thenetwork. The batch size is set to 128 and the polynomialdecay learning rate policy is adopted, with the power pa-rameter set to be 4.

We now summarize the three proposed methods basedon binary codes for presentation attack detection.

• IC-PAD: IrisCodes are used as the input to the PADnetwork. This PAD solution operates on the normal-ized iris region that does not include the pupil re-gion. The normalization results in an IrisCode of size64×256, which is further divided into non-overlappingpatches of size 64 × 64. These sampled patches areused to train the CNN model. This process is indepen-dently applied to the real and imaginary componentsof the IrisCodes, termed Real-IC-PAD and Imag-IC-PAD, respectively. When a test image is presented,the Real-IC-PAD and Imag-IC-PAD average the scoresfrom individual patches. After that, the scores from theReal-IC-PAD and Imag-IC-PAD are averaged to pro-duce the final PA score.

• Pupil-IC-PAD: Pupil-IrisCodes are used as the inputto the PAD network. This PAD solution operates on thenormalized iris region that includes the pupil region.The normalization results in a Pupil-IrisCode of size128 × 256, which is further divided into overlappingpatches of size 128 × 128 with a stride of 32 pixels.The subsequent PA score generation is the same as thatof IC-PAD.

• OC-PAD: OcularCodes are used as the input to thePAD network. This PAD solution operates on un-normalized iris region of size 448× 448. The croppedregions are used to train the CNN model. No patchtessellation is involved here. This process is indepen-dently applied to the real and imaginary componentsof the OcularCode, termed Real-OC-PAD and Imag-OC-PAD, respectively. The PA score of a test image isgenerated by averaging scores from the Real-OC-PADand Imag-OC-PAD.

Both IC-PAD and Pupil-IC-PAD first perform patch-basedscore-level fusion and then the scores from the real andimaginary parts are averaged again; OC-PAD, on the otherhand, fuses scores directly from the real and imaginary partssince it does not involve patch tessellation. Mathematically,the process can be represented as:

S{Re,Im} =1

P

P∑i=1

si. (7)

Here, si is the score of the i-th patch, while SRe and SImare the scores pertaining to the real and imaginary compo-nents. P is the total number of patches, which is set to be

4 and 5 for IC-PAD and Pupil-IC-PAD, respectively. ForOC-PAD, SRe and SIm scores are computed from a sin-gle image, rather than a collection of patches. The final PAscore in this case is computed as 1

2 (SRe + SIm).

4. Experimental ResultsTo verify the effectiveness of the proposed methods, ex-

perimental results are conducted on a publicly available pre-sentation attack dataset and compared against state-of-the-art results.

In order to train the general purpose iris detector in Sec-tion 3.1.2, we used images from the following datasets:LivDet-Iris-2017-Clarkson [28],2 CASIA-Iris-Fake [24],and BERC-Iris-Fake [12]. In total, we used 15,315 iris sam-ples. Here, the iris regions pertaining to both “live”3 and“PA” samples are considered and manually annotated withbounding box positions. The iris detector is trained on thisannotated dataset. Since only a limited number of train-ing samples is available, the iris detection network is fine-tuned on the Darknet-19 model [21] that was pre-trained onthe standard ImageNet dataset. Transfer learning is accom-plished using mini-batch gradient descent with a startinglearning rate of 0.0001, weight decay of 0.0005 and mo-mentum of 0.9. The learning rate is multiplied by 10 afterthe first 100 iterations, and then divided by 10 after 35,000iterations, and stopping at 45,000 iterations. We also utilizemulti-scale training where the network was randomly re-sized to spatial dimensions between 320×320 and 608×608after 10 iterations. The iris detector was used to automati-cally crop both training and test samples for the subsequentiris PAD evaluations. Examples of some iris detection re-sults can be seen in Figure 3.

4.1. Dataset and Evaluation Metrics

This section presents an intra-dataset experiment, whereboth training and test samples originate from the samedataset. The LivDet-Iris-2017-Clarkson dataset was usedfor this evaluation.

Training: The training subset of LivDet-Iris-2017-Clarkson consists of 1,346 print samples, 1,122 patternedcontact samples and 2,469 live samples. The live sampleswere captured using an LG IrisAccess EOU2200 sensor.The PA samples contain printouts of the live images, as wellas 15 different patterned contact lenses.

Testing: The test subset of LivDet-Iris-2017-Clarksonconsists of 908 print samples, 765 patterned contact sam-ples and 1,485 live samples.

We follow the same benchmark protocol and evaluationmetrics as outlined in the 2017 iris liveness detection com-petition [28]: (a) BPCER is the rate of misclassified live

2In this analysis, the training subset of LivDet-Iris-2017-Clarkson isutilized.

3“Live” samples are also referred to as “Bonafide” in the literature.

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 6: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

images (“live” classified as “PA”); and (b) APCER is therate of misclassified PA images (“PA” classified as “live”).In addition, we have reported True Detection Rate (TDR:rate at which PAs are correctly classified) and False Detec-tion Rate (FDR: rate at which live images are misclassifiedas PAs), along with the Receiver Operating Characteristic(ROC) curve. Note that FDR is the same as BPCER andTDR equals (1-APCER).4

4.2. IrisCode versus OcularCode

Both IrisCode and OcularCode share the same encod-ing process, except that the former utilizes the segmentedand normalized iris region. As seen from Figure 5, IC-PADwithout pupil information results in 61.56% TDR at 0.1%FDR. This suggests that it is still feasible to use IrisCodesfor presentation attack detection, although the perfor-mance is not very high. On the other hand, Pupil-IC-PADresults in 79.26% TDR at 0.1% FDR, which is better thanIC-PAD. This clearly reveals that including the pupil infor-mation in the Pupil-IrisCodes improves PAD performance.This underscores the importance of the pupil region, orthe differential between the pupil and iris regions, in de-tecting presentation attacks. Indeed, artifacts from bothprinted and cosmetic contact lens attacks manifest them-selves in the pupil region. Therefore, by including the pupilregion in the normalized image, we expect to see improvedPAD performance, which is observed to be true in this case.

The OC-PAD results in 87.32% TDR at 0.1% FDR. Itmust be noted that OC-PAD operates on the unnormalizediris region. This suggests that the normalization processused to create IrisCodes perhaps discards certain infor-mation that are beneficial for presentation attack detec-tion.

4.3. Analysis by Presentation Attack Type

The iris PAD performance was observed to vary with thenature of the presentation attack. While a presentation at-tack with patterned contact lenses is restricted to the irisregion only, those based on printed images can extend be-yond that region. Hence, the latter is much easier to detect,as evidenced by a higher true detection rate at 0.1% FDR(see Figure 6). This clearly indicates the need to developmore effective PAD solutions for the cosmetic contact PA.

4.4. Comparison Against State-of-the-Art

As shown in Section 4.2, the OC-PAD demonstrated bet-ter performance than the IC-PAD. Hence, we evaluate theproposed OC-PAD against other published results. Threealgorithms that participated in LivDet-Iris 2017 [28] com-

4The LivDet-Iris-2017 competition uses BPCER and APCER as eval-uation metrics. Since FDR and TDR have also been used in the literature,we use these metrics as well.

10−1 100 101 102

False De ec ion Ra e (%)

50

60

70

80

90

100

True

De

ec io

n Ra

e (%

)

IC-PADPupil-IC-PADOC-PAD

Figure 5: Evaluation of IC-PAD and its variants on theLivDet-Iris-2017-Clarkson dataset.

10−1 100 101 102False Detection Rate (%)

50

60

70

80

90

100

True

Detec

tion Ra

te (%

)

w/ Contactw/ PrintFull

Figure 6: Evaluation of OC-PAD performances on theLivDet-Iris-2017-Clarkson dataset with respect to differentpresentation attack types.

petition were used for comparison. “ANON1” algorithm5

achieved 33.46% and 0.44% APCERs on patterned con-tact lens and printed attacks, respectively. This is consis-tent with our observation made in Section 4.3. Next, weperform an ablation study to investigate the performanceof OC-PAD with respect to the real and imaginary compo-nents. As can be seen in Table 2, both Real-OC-PAD andImag-OC-PAD independently perform better than state-of-the-art methods at different BPCER values. OC-PAD aver-ages the PA scores from Real-OC-PAD and Imag-OC-PAD,and it demonstrates better performance than the individual

5The other two participating algorithms were not evaluated with respectto different presentation attack types on this dataset.

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 7: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

10−1 100 101 102False Detectio Rate (%)

60

65

70

75

80

85

90

95

100

Tr%e Detectio

Ra

te (%

)

MT-PAD [2]Iris-PADReal-OC-PADImag-OC-PADOC-PAD

Figure 7: Evaluation of intra-dataset iris PAD performanceon the LivDet-Iris-2017-Clarkson dataset. Note that theOC-PAD is only marginally inferior to that of Iris-PAD,which utilizes the raw pixel values rather than binary codes.

components. This suggests that these two components pro-vide complementary information.

Next, we also plot the ROC curves for individual meth-ods in Figure 7. MT-PAD [2] is a multi-task solution pre-viously proposed in the literature that simultaneously per-forms iris detection and presentation attack detection. Iris-PAD was implemented by directly using the raw pixel inten-sities rather than the corresponding binary code. Comparedto Iris-PAD, the performance of OC-PAD is slightly lower.This could be due to the extensive data augmentation usedby Iris-PAD during training, while the Real-OC-PAD andImag-OC-PAD use limited data augmentation, such as crop-ping and rotation only. Since we operate on binary codes,the color jittering augmentation scheme degrades PAD per-formance and is not used as an augmentation scheme in thiscase. This suggests that not all existing data augmentationmethods are effective for binary codes. In this regard, weneed to design data augmentation methods that are tailoredto the binary codes. The simple sum-fusion rule has demon-strated superior performance for OC-PAD. More complexfusion schemes, such as two-stream CNN [23], could beexplored in the future.

4.5. Feature Visualizations from OcularCode

To understand the features learned by the proposed irisPADs, we visualize the activation layers of the network dur-ing the forward pass. Here, the Imag-OC-PAD networktrained using the LivDet-Iris-2017-Clarkson dataset is usedas an example to interpret features learned by the CNN. Wevisualize the activation maps when an input cropped irissample is encoded with the imaginary bits of the Ocular-

Code. The first convolution layer outputs a 448× 448× 32tensor. We slice this tensor into 32 different feature maps,corresponding to 32 different filters. We then visualize afew of these feature maps. Figure 8 shows these featuremaps when imaginary bits of the OcularCode calculatedfrom both a live and a print sample are fed into the network.

Figure 8: Visualization of selected feature maps from thefirst convolutional layer of the Imag-OC-PAD for both a live(first two rows) and a print sample (bottom two rows).

We also visualize the final convolutional layer featuresusing the t-SNE algorithm [27], which reduces the featuremap dimension from 392 to 2 (see Figure 9). The embed-ding representations learned from the CNN model show thatPA samples can be separated from live samples with goodaccuracy, which is consistent with our previous experimen-tal results reported in Table 2. However, at the same time,the learned model shows that PA samples are more likely tobe classified as live samples.

5. Conclusions

This paper investigated the feasibility of using IrisCodesfor presentation attack detection. In general, we studied theefficacy of using binarized Gabor phasor features for thistask. Our analysis suggests that incorporating pupil infor-mation in the IrisCode is beneficial for presentation attackdetection. In addition, extracting binarized Gabor phasorfeatures from the un-normalized iris is observed to furtherimprove detection accuracy. The proposed approach could

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 8: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

Table 2: Comparison of the proposed OC-PAD against other state-of-the-art CNN results (%). Note that BPCER is the sameas FDR and APCER is the same as (1-TDR).

APCER (%)BPCER (%) CASIA [28] ANON1 [28] UNINA [28] Real-OC-PAD Imag-OC-PAD OC-PAD5.65 9.61 - - 8.31 7.71 7.533.64 - 15.54 - 9.09 8.73 8.130.81 - - 13.39 12.55 13.03 11.0

-100 -50 0 50 100

First Dimension

-150

-100

-50

0

50

100

Second D

imensio

n

Live

PA

Figure 9: t-SNE feature visualizations on the LivDet-Iris-2017-Clarkson test subset using Imag-OC-PAD.

be potentially combined with iris recognition systems to of-fer a unified solution for both iris recognition and iris pre-sentation attack detection. Future work will involve deter-mining the features that have been learned by the proposedCNNs from the three different types of binary codes. Uti-lizing different bandwidth values for the Gabor filters maybe essential to tease out binary phasor codes that are moreeffective for the problem of presentation attack detection.

AcknowledgementThis research is based upon work supported in part by the Of-

fice of the Director of National Intelligence (ODNI), IntelligenceAdvanced Research Projects Activity (IARPA), via contract num-ber 2017-17020200004. The views and conclusions containedherein are those of the authors and should not be interpreted asnecessarily representing the official policies, either expressed orimplied, of ODNI, IARPA, or the U.S. Government. The U.S.Government is authorized to reproduce and distribute reprints forgovernmental purposes notwithstanding any copyright annotationtherein.

References[1] S. Adamovic, M. Milosavljevic, M. Veinovic, M. Sarac, and

A. Jevremovic. Fuzzy commitment scheme for generation ofcryptographic keys based on iris biometrics. IET Biometrics,6(2):89–96, 2017.

[2] C. Chen and A. Ross. A multi-task convolutional neural net-work for joint iris detection and presentation attack detec-tion. In Proc. of WACV Workshops, 2018.

[3] A. Czajka. Pupil dynamics for iris liveness detection.IEEE Transactions on Information Forensics and Security,10(4):726–735, 2015.

[4] A. Czajka and K. W. Bowyer. Presentation attack detectionfor iris recognition: An assessment of the state of the art.CoRR, 2018.

[5] J. Daugman. How iris recognition works. IEEE Transactionson Circuits and Systems for Video Technology, 14(1):21–30,2004.

[6] J. Galbally, A. Ross, M. Gomez-Barrero, J. Fierrez, andJ. Ortega-Garcia. Iris image reconstruction from binary tem-plates: An efficient probabilistic approach based on geneticalgorithms. Computer Vision and Image Understanding,117(10):1512–1525, 2013.

[7] L. He, H. Li, F. Liu, N. Liu, Z. Sun, and Z. He. Multi-patch convolution neural network for iris liveness detection.In Proc. of BTAS, 2016.

[8] Z. He, Z. Sun, T. Tan, and Z. Wei. Efficient iris spoof de-tection via boosted local binary patterns. In Proc. of ICB,2009.

[9] S. Hoffman, R. Sharma, and A. Ross. Convolutional neuralnetworks for iris presentation attack detection: Toward cross-dataset and cross-sensor generalization. In Proc. of CVPRWorkshops, 2018.

[10] N. Kohli, D. Yadav, M. Vatsa, R. Singh, and A. Noore. Syn-thetic iris presentation attack using iDCGAN. In Proc. ofIJCB, 2017.

[11] O. V. Komogortsev, A. Karpov, and C. D. Holland. Attackof mechanical replicas: Liveness detection with eye move-ments. IEEE Transactions on Information Forensics and Se-curity, 10(4):716–725, 2015.

[12] S. Lee, K. Park, Y. Lee, K. Bae, and J. Kim. Multifeature-based fake iris detection method. Optical Engineering,46(12), 2007.

[13] L. Masek. Recognition of human iris patterns for biomet-ric identification, Bachelor’s Thesis, University of WesternAustralia, 2003.

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.

Page 9: Exploring the Use of IrisCodes for Presentation …rossarun/pubs/ChenRossIrisCodes...un-normalized iris is used). 3.1.1 Iris Segmentation and Normalization In this approach, we first

[14] R. Mukherjee and A. Ross. Indexing iris images. In Proc. ofICPR, 2008.

[15] K. Nandakumar and A. K. Jain. Multibiometric templatesecurity using fuzzy vault. In Proc. of BTAS, 2008.

[16] K. Nguyen, C. Fookes, R. Jillela, S. Sridharan, andA. Ross. Long range iris recognition. Pattern Recognition,72(C):123–143, 2017.

[17] K. Nguyen, C. Fookes, A. Ross, and S. Sridharan. Iris recog-nition with Off-the-Shelf CNN features: A deep learning per-spective. IEEE Access, 6:18848–18855, 2018.

[18] F. Pala and B. Bhanu. Iris liveness detection by relative dis-tance comparisons. In Proc. of CVPR Workshops, 2017.

[19] R. Raghavendra and C. Busch. Robust scheme for iris pre-sentation attack detection using multiscale binarized statisti-cal image features. IEEE Transactions on Information Foren-sics and Security, 10(4):703–715, 2015.

[20] R. Raghavendra, K. B. Raja, and C. Busch. ContlensNet:Robust iris contact lens detection using deep convolutionalneural networks. In Proc. of WACV, 2017.

[21] J. Redmon and A. Farhadi. YOLO9000: better, faster,stronger. In Proc. of CVPR, 2017.

[22] I. Rigas and O. V. Komogortsev. Eye movement-driven de-fense against iris print-attacks. Pattern Recognition Letters,68:316–326, 2015.

[23] K. Simonyan and A. Zisserman. Two-stream convolutionalnetworks for action recognition in videos. In Proc. of NIPS,pages 568–576, 2014.

[24] Z. Sun, H. Zhang, T. Tan, and J. Wang. Iris image clas-sification based on hierarchical visual codebook. IEEETransactions on Pattern Analysis and Machine Intelligence,36(6):1120–1133, 2014.

[25] C.-W. Tan and A. Kumar. Efficient and accurate at-a-distanceiris recognition using geometric key-based iris encoding.IEEE Transactions on Information Forensics and Security,9(9):1518–1526, 2014.

[26] J. E. Tapia, C. A. Perez, and K. W. Bowyer. Genderclassification from the same iris code used for recognition.IEEE Transactions on Information Forensics and Security,11(8):1760–1770, 2016.

[27] L. van der Maaten and G. Hinton. Visualizing data using t-SNE. Journal of Machine Learning Research, 9:2579–2605,2008.

[28] D. Yambay, B. Becker, N. Kohli, D. Yadav, A. Czajka,K. W. Bowyer, S. Schuckers, R. Singh, M. Vatsa, A. Noore,D. Gragnaniello, C. Sansone, L. Verdoliva, L. He, Y. Ru,H. Li, N. Liu, Z. Sun, and T. Tan. LivDet Iris 2017 - irisliveness detection competition 2017. In Proc. of IJCB, 2017.

[29] Z. Yan, L. He, M. Zhang, Z. Sun, and T. Tan. Hierarchicalmulti-class iris classification for liveness detection. In Proc.of ICB, 2018.

C. Chen and A. Ross, "Exploring the Use of IrisCodes for Presentation Attack Detection," Proc. of 9th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS),

(Los Angeles, USA), October 2018.