Professional Documents
Culture Documents
16, 2021
Abstract— Iris recognition has emerged as one of the most (COVID19) pandemic. Iris image acquisition modules widely
accurate and convenient biometric for person identification use near-infrared (NIR) illumination, typically in the wave-
and has been increasingly employed in a wide range of e- length range of 700-900nm, which can reveal enhanced quality
security applications. The quality of iris images acquired at-
a-distance or under less constrained imaging environments is of iris texture under constrained imaging environment. How-
known to degrade the iris recognition accuracy. The periocular ever, with the increase in the standoff distances, the quality of
information is inherently embedded in such iris images and can acquired iris images significantly degrades. In such imaging
be exploited to assist in the iris recognition under such non- scenarios, the periocular information can play an increasingly
ideal scenarios. Our analysis of such iris templates also indicates important role for accurate personal identification. In recent
significant degradation and reduction in the region of interest,
where the iris recognition can benefit from a similarity distance years, periocular recognition has been receiving increasing
that can consider importance of different binary bits, instead of attention for its promising performance under such less con-
the direct use of Hamming distance in the literature. Periocular strained imaging conditions [3]- [4]. The periocular region
information can be dynamically reinforced, by incorporating the usually refers to the region around the eye, which preferably
differences in the effective area of available iris regions, for more includes the eyebrow [5]. Such periocular near-infrared iris
accurate iris recognition. This article presents such a periocular-
assisted dynamic framework for more accurate less-constrained images, in particular, present highly discriminative features
iris recognition. The effectiveness of this framework is evaluated for the person identification. Earlier work [3], [4], [6], [7], [8]
on three publicly available iris databases using within-dataset in this area has validated that the periocular region is highly
and cross-dataset performance evaluation, e.g., improvement in discriminative among different persons, and can be considered
the recognition accuracy of 22.9%, 10.4% and 14.6% on three as an effective alternative or supplement to the face or iris
databases under both the verification and recognition scenarios.
recognition especially when the entire face or clear iris images
Index Terms— Iris recognition, periocular recognition, per- are not available. This work is motivated to further such
sonal identification, biometrics, deep learning. advances in the less-constrained iris recognition capabilities
and introduces a new framework to more accurately and
I. I NTRODUCTION adaptively match less-constrained iris images.
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 867
2D discrete Fourier transforms (DFT) which offers promising iris images that are represented in a segmentation-less polar
results. In 2009, Sun and Tan [12] employed the multi-lobe domain.
differential filter (MLDF) referred to as the ordinal filters 2) Periocular Recognition: In recent years, researchers
which offer an alternative for the Gabor/log-Gabor filters in have devoted consistent efforts to investigate new periocular
generating rich iris feature templates. recognition algorithms for the images acquired under less-
Iris recognition research has also attracted a variety of constrained environments [7], [24]. Earlier feasibility study
approaches to enhance segmentation accuracy for the acquired on using the periocular regions for human recognition under
iris images and accurate segmentation is critical in enhanc- varying imaging conditions was undertaken by Park et al. [6]
ing reliability for the iris recognition. Some of the most in 2009, and promising results were reported, which provided
widely employed iris segmentation algorithms are based on support for the subsequent research. Bharadwaj et al. [4]
the integro-differential operator [1] and circular Hough trans- further explored the effectiveness of periocular recognition in
forms [9] which are adapting for detecting iris and pupil situations arising from the failure of iris recognition. In this
circles from the near-infrared eye images. These methods work, part of the later research focuses on cross-spectral
perform well for the high-quality iris images but are quite periocular matching [25] using the potential from the neural
known to be least reliable for the noisy images acquired networks. The above explorative works have further motivated
under relaxed environments. Tan et al. [13] proposed an the researchers to continuously improve the matching accuracy
iterative approach to coarsely cluster the iris and non-iris of periocular images. In 2013, another promising approach
region pixels before applying the integro-differential operator, appeared in [3], which exploits key-point features and spatial-
and achieved higher reliability in segmenting the iris pixels filter banks, i.e., Dense-SIFT and LMF features, followed by
from noisy iris images. Following the similar coarse-to-fine K-means clustering for dictionary learning and representa-
strategy, a competitive approach was detailed in [3] which tion. However, this approach does not investigate periocular-
makes use of the random walker algorithm [14] for coarsely specific feature representations, and the uses computationally
locating the iris region, followed by a couple of gray-level demanding Dense-SIFT features matching. Smereka et al.
statistics based operators to refine the boundaries. These oper- [5] proposed the periocular probabilistic deformation model
ators have shown to enable pixel-level precision in the final (PPDM) in 2015, which provides sound modeling for the
output or the iris masks. More recent approach [15] utilized potential deformations that exists between two matched peri-
an improved total variation model to address accompanying ocular images. Inference of the captured deformation using
noise and artifacts in less constrained iris images. [16] relied correlation filter is utilized for matching periocular image
on the color/illumination correction along with the watershed pairs. Later in 2016, the same group of researchers improved
transform to segment noisy iris images acquired under visible their basic model by selecting discriminative patch regions for
wavelength. more accountable matching [26]. These two methods achieve
There is quite limited work to exploit the potential of deep promising performance on multiple datasets. Nevertheless,
neural network capabilities for the iris recognition, especially both of them rely on patch-based matching scheme, and there-
while considering the tremendous popularity of deep learning fore are more susceptible to scale variations or misalignment,
for various computer vision tasks such as face recognition [17]. that often violate the patch correspondences, which is more
An earlier attempt for deep representation of iris appeared in likely to happen during the real deployments.
[18] in 2015, but such proposal was to detect presentation Deep learning techniques, especially convolutional neural
attacks, a two-class classification problem, instead of the iris networks (CNN), have gained immense popularity for com-
recognition. A new approach using DeepIrisNet was inves- puter vision and pattern analysis tasks in recent years. The
tigated in [19] and used a deep learning-based framework recent survey on periocular recognition methods [7], [24]
for general iris recognition. This work is essentially a direct suggests that few studies have considered the potential from
application of typical convolutional neural networks (CNN) deep learning techniques to boost the periocular matching
without many optimizations for the iris patterns. Another more accuracy. Reference [27] provided insightful observations on
recent work in [20] has attempted to exploit a deep belief periocular features and comparison of machine with human
net (DBN) for iris recognition. Its core component is the matching performance. In [28], Bowyer and Burge presented
optimal Gabor filter selection to generate IrisCode, then they a systematic summary on the related ocular recognition sys-
feed the generated IrisCode to DBN to extract deep learning tems and algorithms. More recently, Proença and Neves [29]
features without any iris-specific optimization. More recent claimed that iris and sclera regions might be less reliable for
work in [21] proposed a UniNet [21] employing the deep periocular recognition and proposed Deep-PRWIS. In their
fully convolutional networks (FCN) [22] to generate iris binary work, periocular images are augmented with inconsistent iris
images and masks for Hamming distance calculation, which and sclera regions for training a deep CNN, so that the network
explores the substantial connections between iris recognition implicitly degrades the iris and sclera features during learning.
and deep learning. This work introduces a new loss func- Promising results were reported from the Deep-PRWIS on
tion that incorporates conventional bit-shifting operations and two public databases. More promising efforts appeared in
masks in matching score computations, and achieves state-of- [8], which uses a deep learning-based architecture for robust
the-art accuracy on several publicly available datasets. Another and accurate periocular recognition incorporating the attention
related and promising work appeared in [23] which uses a deep model to emphasize the region with higher discriminative
learning architecture to infer misalignment between a pair of information. This algorithm achieves state-of-the-art accuracy
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
868 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 16, 2021
on six publicly available databases and can serve as a reason- motivated to address such iris recognition challenges and
able baseline for further research in this area. evaluate iris recognition capabilities under more realistic sce-
3) Iris and Periocular Feature Fusion: The periocular infor- narios. Iris images acquired under less constrained imaging
mation is simultaneously accessible from the iris images and environments often present varying regions of effective iris
therefore its use to achieve better iris recognition performance pixels [1], [40]. In the context of such less constrained
is a feasible strategy. Quite a few prior works have attracted iris images, we revisit the conventional Hamming distance
attention to this aspect and several approaches have non-ideal to match binarized iris templates. Such iris images present
scenarios. In 2010, Woodard et al. [33] combined iris and significant variations in occlusions which should be carefully
periocular features using score level combination, i.e., weight considered while simultaneously utilizing available periocular
sum rule, to improve the recognition performance in non- features. Since iris information is inherently embedded in peri-
ideal iris imagery. Optimal weights for the two modalities are ocular images, the effectiveness of iris matching can benefit
empirically obtained. Another promising attempt appeared in from the relative attention or eye area, like for the human
[3] which simultaneously recovers the iris feature extracted visual systems [41]. Table I presents a summary of related
from log-Gabor filters, and periocular features extracted from work with our work in this article for the less constrained
Dense-SIFT and LMF, to enhance the iris recognition accuracy iris recognition. The key contributions from this article can be
under relaxed imaging constraints. Raja et al. [34] proposed summarized as follows:
a framework to combine the information from face, iris and • This article introduces a new framework for the perioc-
periocular biometric modalities for the user authentication on ular assisted iris recognition. Iris images under a less-
their smartphones. Various score level combination schemes constrained imaging environment often present varying
are explored, including min rule, max rule, product rule, regions of effective iris pixels for the iris matching. Such
and weighted- score fusion rule, where the weight for each differences in the effective number of available iris pixels
modality is determined according to its contribution to the can be used to dynamically reinforce periocular infor-
recognition performance. Besides, some approaches adopt mation which is simultaneously available from such iris
learning-based score-level fusion strategies. Santos et al. [35] images. Such dynamic reinforcement should also consider
presented an artificial neural network with two hidden layers effective regions of discriminative features that receive
to fuse iris and periocular information at the score level for the varying attention during respective periocular matching.
mobile cross-sensor applications. Verma et al. [36] utilized the Our framework therefore incorporates such discriminative
random decision forest (RDF), which is an ensemble learning information using a multilayer perceptron network for the
method, to combine the match scores of iris and periocular less-constrained iris recognition. The experimental results
biometrics. Noticeable improvement in the performance is presented in Section III-B for within-database match-
shown for at-a-distance person recognition. Ahuja et al. [37] ing using the receiver operating characteristics curve
extracted the periocular feature using deep learning and the iris (ROC) on three publicly available databases, indicate
feature using the root SIFT. Then they combined the match outperforming results over state-of-the-art methods. Also,
scores from these two modalities using the mean rule and the ROC results presented in Section III-C show that our
linear regression. There are some other promising attempts algorithm outperform others in cross-dataset matching.
in the literature that integrate the information from these two The results from within dataset matching and cross-
biometric at the decision level and feature level combination. dataset matching validate the effectiveness and general-
Talreja et al. [38] fused iris and periocular modality at the ization ability of the framework presented in this article
decision level to increase the reliability in the unconstrained for the less-constrained iris recognition.
iris recognition. They trained a logistic regression model to • The importance of black (0) and white (1) pixels in bina-
predict the weights for each of the classifiers and obtain a rized iris templates may not be the same or similar for iris
final response. Joshi et al. [39] investigated iris and periocular image templates acquired under less constrained imaging.
biometric performance from their feature level combination. Therefore, this article presents a new approach to match
They first concatenate iris and periocular features and then such templates using a similarity measure, instead of
employ the direct linear discriminant analysis (DLDA) to Hamming distance in the literature, which can accom-
obtain discriminative and low dimensional feature vectors modate the importance of different bits in iris templates.
for the final classification. More recently, Zhang et al. [30] The experimental results presented in this article on three
provided a promising framework to combine the iris and publicly available iris databases consistently indicate out-
periocular features extracted from maxout CNN to enhance performing results and validate the effectiveness of such
the performance for mobile based personal identification. approach for less-constrained iris recognition.
• Iris recognition, unlike for verification, is required
B. Our Work in a range of security or civilian applications, e.g.,
Accuracy of iris recognition under a less-constrained envi- in search of missing children or wanted suspects. Most
ronment is known to significantly degrade, as compared to state-of-the-art methods in the literature [8], [21], [30]
those from the conventional or standoff iris recognition sys- have only focused on verification problems and not
tems. Such iris images are generally acquired with greater evaluated capabilities for the recognition problem.
standoff distances, for the surveillance or from the mobile This article presents iris recognition performance on
devices with less-cooperative individuals. This research is three public databases, using both open set and closed
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 869
TABLE I
C OMPARATIVE S UMMARY OF R ELATED AND R ECENT W ORK ON L ESS -C ONSTRAINED I RIS R ECOGNITION
set protocols. The experimental results presented in [21] to achieve accurate iris matching while the AttenNet and
Section III-C of this article, using average rank-one FCN-Peri [8] are embedded in simultaneously matching the
accuracy from the cumulative matching characteristics periocular regions in the acquired eye or iris images. The
(CMC), achieve outperforming results over the other network is trained during two different training or offline
state-of-the-art methods in the literature. We also present phases. We firstly pre-process each of the acquired eye images
open-set and cross-distance experimental results in to independently recover the normalized iris images from
Section III-D and III-E, unlike in the literature, which respective periocular images. The corresponding region of
further validate the effectiveness of proposed approach. interest images is fed to the respective subnets and trained
Comparative performance from our approach with other independently during the first network training phase. During
competing methods, on three common and public iris images the second training phase, all the parameters in two subnets
datasets, is also summarized in Table I. The rest of this article are frozen and used to recover several cues that indicate the
is organized as follows. Section II provides details on our similarity among the iris and periocular templates, including
unified framework for less constrained iris recognition. This the effective region of iris images among matched template
section also includes the architectures for iris and periocular and the corresponding periocular region components among
recognition, together with the formulation of the dynamic the matched templates. Finally these cues from the two subnets
fusion approach introduced in this work. Our comparative are employed to train a multilayer perceptron (MLP) network
experimental results from within dataset matching and that can enable a binary prediction using the softmax cross-
cross-dataset matching using three different public databases entropy loss. During the performance evaluation or the test
are presented in Section III, The discussion section appears phase, a pair of eye images are fed into the trained models
in Section IV which also includes comparative results from and recover the prediction results from the last softmax layer.
other schemes. The key conclusions from this article are These softmax layer results are considered as the consolidated
summarized in Section V. match scores between the input or the unknown eye pair
images. These consolidated match scores are used to achieve
II. P ERIOCULAR -A SSISTED I RIS R ECOGNITION the binary classification decisions for different applications.
F RAMEWORK Following sections provide further details on different compo-
The framework for periocular-assisted and multi-feature nents of the framework.
collaboration schemes to achieve dynamic iris recognition is
illustrated in Figure 1. The detailed explanation of different A. Iris Template Generation and Comparisons
blocks in this diagram is systematically introduced in the Each of the acquired eye images is first subjected to the
following three sections. This framework adopts the UniNet localization of region of interest or the iris segmentation and
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
870 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 16, 2021
Fig. 1. The framework for the deep dynamic fusion using iris and periocular information.
image normalization. The preprocessing step to normalize iris where N is the batch size, M A ,M P ,M N are the corresponding
images are same as the earlier work in [15]. The dimension of masked feature maps generated by the UniNet, and m is
all the segmented and normalized iris images generated from the hyperparameter controlling the margin between anchor-
the preprocessing steps, from all the databases employed in positive and anchor-negative distances which is set to be 0.1 in
our work, is 512 × 64 pixels. After the normalization, these our experiments.
images are also subjected to the contrast enhancement which
saturates 5% of iris region pixels at high and low intensities. B. Comparisons Using Similarity Score
The normalized rectangular iris images are subjected to
recover respective feature templates and respective masks Hamming distance is widely employed to compute the
depicting valid iris pixels or regions. The UniNet architecture dissimilarities between two binary feature templates in a range
introduced in [21] has shown to offer state-of-the-art iris of biometric identification problems, such as for the iris or the
matching capabilities and was also adopted in this work. palmprint recognition. It assumes that the information content
The UniNet includes two fully-convolutional sub-networks from all the template values in the coding space is equally
called FeatNet and MaskNet as specified in Table II. The important to distinguish the user identity. However, the choices
MaskNet generates binary mask distinguishing the valid and of feature extraction and binarization methods, along with
invalid or less reliable regions in the iris templates that often the nature of input images, can effectively determine the
degrade the iris matching accuracy. The UniNet uses triplet importance of white (ones) area and black (zeros) area in the
architecture for the training. The triplet input includes one encoded images. Therefore, a more flexible distance measure
anchor image (A), one positive image (P) and one negative that can consider such asymmetric importance is proposed
image (N). Anchor-positive is a genuine pair of images which to be incorporated for matching less-constrained iris images.
are different samples from the same class, while anchor- Such measure is also referred to as the weighted similarity
negative is an imposter pair of images which are samples score (WS) with azzoo similarity measure [43] and can also
from different classes. In this work, we generate triplets in a be incorporated in matching iris templates.
ratio of 1:3 between the genuine match pairs and the imposter The effectiveness of white pixel matching and the black
match pairs for the respective training sets since the number pixel matching in feature templates can also be experimentally
of imposter pairs are much more than the number of genuine evaluated. Let us assume that the number of white pixels and
pairs in any datasets. The MaskNet is pre-trained using ND- black pixels from one feature template A can be respectively
IRIS-0405 Iris Image Dataset [42] and all the parameters are represented as PW (A) and PB (A). While comparing template
frozen in this work. The FeatNet, pretrained with ND-IRIS- A with template B, we can perform only white pixels matching
0405 Iris Image Dataset and publicly made available from MW (A, B) and only black pixels matching M B (A, B), then
[21], is finetuned using the triplet pairs generated from the we can compute the white pixel matching rates Rw (A, B) and
respective training sets. The FeatNet is essentially a fully black pixel matching rate R B (A, B) as shown in the following
convolutional neural network and aims to learn the same size two equations.
but more robust pseudo-binary representation of the input iris 2 × MW (A, B)
RW (A, B) = (2)
images. The loss function introduced in the FeatNet training PW (A) + PW (B)
is the extended triplet loss which aims to enlarge the margin 2 × M B (A, B)
of the pseudo-Hamming distance between the intra-class and R B (A, B) = (3)
PB (A) + PB (B)
inter-class matching. The extended triplet loss L E T L can be
defined as follows. The difference in the contributions from different pixels match-
ing,i.e. average RW and R B from the genuine and imposter
pairs, can also be empirically observed from the experiments
1
N
L ET L = max(||M P − M A ||2 −||M N − M A ||2 +m, 0) using templates generated from the databases. We select
N 1,000 genuine matching and 2,000 imposter matching from the
i=1
(1) test on CASIA-Mobile-V1-S3 dataset for empirical evaluation.
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 871
It was observed that the average RW are 0.5733 and the TABLE II
average R B is 0.6138 for the genuine matches, while RW are T HE S PECIFICATION OF I NCORPORATED UniNet
0.4159 and the average R B is 0.4563 for the imposter matches.
In order to accommodate differences in the discriminative
information from white pixel pairs and the black pixel pairs,
we use different weight and generate weighted similarity
measure as follows:
⎧
⎨2 − α i f Ii, j = Ii, j = 1
⎪ 1 2
1 2
W S(Ii, j , Ii, j ) = α i f Ii,1 j = Ii,2 j = 0 (4)
⎪
⎩
0 i f Ii, j = Ii, j
1 2
where Ii,1 j ,Ii,2 j are pixels in row i and column j in two matched
two iris templates, and α is hyperparameter controlling the
significance of coding pairs. In all our experiments, α is set
as 0.3. Assuming the image size of iris images are H × W ,
we generate the match score using the weighted similarity as
follows:
1 H W
SW S = W S(Ii,1 j , Ii,2 j ) (5)
H ×W
i=1 j =1
It can be observed that when the α is unity, the value of Sws
is essentially the difference between unity and the normalized
Hamming distance. Therefore, weighted similarity can be
considered as a more flexible alternative for the iris templates
matching.
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
872 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 16, 2021
generate more reliable consolidated match score between two III. E XPERIMENTS AND R ESULTS
unknown or input images. We perform a series of experiments on three publicly
As illustrated in Figure 1, the UniNet generates pseudo- available datasets to ascertain the effectiveness of the proposed
binary feature maps, along with the respective masks, while the framework for less-constrained iris recognition. This section
AttenNet generates the feature vectors to compute Euclidean firstly provides brief but necessary information for the three
distance among respective ROI maps. Therefore, we can simul- public datasets used in this work. We then explain the exper-
taneously generate iris match scores and periocular match imental protocols in the following section. This section also
score using the Euclidean distance. Another important input provides a comparative analysis of results from our method
for MLP, which effectively represents the importance or the with other state-of-the-art methods.
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 873
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
874 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 16, 2021
Fig. 4. Comparative ROC results from cross-dataset performance evaluation. The results in (a)-(b) use the model trained with CASIA.v4 Distance dataset
while results in (c)-(d) use the model trained with CASIA-Mobile-V1-S3 dataset.
TABLE V
S UMMARY OF R ECOGNITION R ATES AND E QUAL E RROR R ATES VALUES F ROM THE C OMPARISON W ITHIN D ATASET M ATCHING
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 875
TABLE VII
S UMMARY OF I DENTIFICATION R ATES AND E QUAL E RROR R ATES VALUES FOR THE open set P ERFORMANCE E VALUATION
Fig. 5. Comparative CMC results from cross-dataset matching. (a)(b) use the model trained with CASIA.v4 Distance dataset. (c)(d) use the model trained
with CASIA-Mobile-V1-S3 dataset.
aimed to validate the generalization capability of our frame- children, requires open set matching capabilities under cross-
work, especially when the image samples available for the dataset scenarios. Therefore, we also perform open-set perfor-
training are quite limited. The EER values are summa- mance evaluation, using the cross-database performance eval-
rized in Table VI and the respective ROCs are shown in uation configuration in Section III-C, to further ascertain the
Figure 4. effectiveness of our framework. The closed-set identification
The results summarized in this set of experiments indi- presented in Section III-B and III-C performs comparisons
cate consistent improvement from our framework during the between the gallery and the known probes. The open-set
cross-database matching which reveals the generality of the identification includes known subjects that enrolled in the
framework in matching less-constrained iris images. We also gallery and unknown subjects which are only seen during
perform additional cross-database experiments for the recog- the query time. The open-set identification objective is to
nition problem and use the cumulative matching characteristic correctly identify probe subjects that are present in the gallery
(CMC) curve to ascertain such performance in Figure 5. while rejecting all other probe queries as unknown, and it
In such experiment, we follow the leave-out-one protocol is widely applied in surveillance and person reidentification
which means that we select one probe images from the test set applications. In our experiments, we use the last 30 test
and perform matching with other images. These results also subjects as unknown subjects in all or respective datasets,
present encouraging and superior performance, to a varying which are not enrolled in the gallery. We then ascertain the
degree, and validate generalization of our approach for the open-set performance using the false positive identification
related iris recognition problem. rate (FPIR) against false negative identification rate (FNIR),
similar to other references, e.g., [47]. These plots are shown
D. Open Set Performance Evaluation in Figure 6 while their identification rates and equal error
The deployment of less-constrained iris recognition algo- rates are summarized in Table VII. These results indicate
rithms, for applications in surveillance or in search of missing consistent performance improvement, to varying degrees, and
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
876 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 16, 2021
Fig. 6. Comparative FPIR against FNIR results from the open set performance evaluation. (a)(b) use the model trained with CASIA.v4 Distance dataset.
(c)(d) use the model trained with CASIA-Mobile-V1-S3 dataset.
TABLE VIII
S UMMARY OF C OMPARATIVE EER VALUES FOR THE cross-distance P ER -
FORMANCE E VALUATION
TABLE IX
S UMMARY OF E QUAL E RROR R ATES VALUES F ROM THE C OMPETING
F USION M ETHODS
underline merit of our framework for the cross-database open- Fig. 7. Comparative ROC results from the cross-distance experiments for
set recognition applications. the same subjects.
E. Cross Distance Performance Evaluation in matching the iris images from different distances when
Cross-distance performance evaluation can allow us to the model or training data is only available from some other
ascertain the effectiveness of the iris recognition framework (generally nearer) distances. Such performance can provide
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 877
Fig. 9. Comparative receiver characteristic curve (ROC) results from different fusion algorithms in within dataset matching.
us insights on actual performance expected during the less performance degrades under cross-distance matching and is
constrained deployment when enrolled subjects from one quite expected. However, significant improvement observed in
distance are matched with those subjects imaged at some the matching accuracy, using our approach, can help to validate
other (higher) distances. Therefore, in this set of experi- the effectiveness of our framework for the cross-distance iris
ments, we use the model trained from iris images acquired matching.
from 5 feet distance, i.e. using Q-FIRE-05-middle-illumination
dataset. This model is used to match same subjects iris IV. D ISCUSSION
images acquired from 7 feet distance (using the Q-FIRE- The complementary nature of match scores generated from
07-middle-illumination dataset) and 11 feet distance (using the deep features in our experiments can be visualized from
Q-FIRE-11-middle-illumination dataset). These images are the two-dimensional plots representing iris and periocular
collected from 7 feet and 11 feet respectively under the scores. Figure 8 illustrates such plots for the distribution
same illumination condition and therefore represent more of (normalized) genuine and imposter scores from iris and
realistic scenarios for less-constrained iris recognition. The periocular matching using respective databases. The subplots
4,703 samples from 159 right eyes in Q-FIRE-05-middle- in each axis are the kernel density estimation of each score
illumination dataset compose the gallery, and attempt to use distribution. These plots from less-constrained images indicate
10 samples from same eye as the test probe to match with that the joint use of individual match scores can be used to
respective gallery. There are 1,576 test samples from Q-FIRE- more effectively separate genuine and impostor match scores
07-middle-illumination dataset and 1,543 test samples from as pursued in this work.
the Q-FIRE-11-middle-illumination dataset. These matchings We also attempt to incorporate other popular fusion methods
ensure that every genuine (also impostor) match pairs use to achieve performance improvement. The weighted sum has
two images that are acquired from different distances. After been emerged as more popular fusion rule and was also
performing such matching, we get 46,110 genuine match attempted in our experiments. We also used min (best of two-
scores and 7,365,318, imposter match scores for Q-FIRE-07- match score), product, and nonlinear fusion [48] which has
middle-illumination dataset, and 45,665 genuine match scores shown to offer superior performance over weighted sum in
and 7,215,767 imposter match scores for Q-FIRE-11-middle- other biometrics applications. Such comparative experimental
illumination dataset. The ROC results are shown in Figure 7 results from different databases are presented in Figure 9 and
while respective EER values are summarized in Table VIII. the respective EER values are summarized in Table IX. These
It can be observed from these results that the average matching results indicate the effectiveness of dynamic fusion approach
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
878 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 16, 2021
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.
WANG AND KUMAR: PERIOCULAR-ASSISTED MULTI-FEATURE COLLABORATION 879
[27] K. P. Hollingsworth, S. S. Darnell, P. E. Miller, D. L. Woodard, [44] R. Girshick, “Fast R-CNN,” in Proc. IEEE Int. Conf. Comput. Vis.,
K. W. Bowyer, and P. J. Flynn, “Human and machine performance on Dec. 2015, pp. 1440–1448.
periocular biometrics under near-infrared light and visible light,” IEEE [45] OpenCV Based Face and Eye Detector. Accessed: Jan. 2018.
Trans. Inf. Forensics Security, vol. 7, no. 2, pp. 588–601, Apr. 2012. [Online]. Available: http://docs.opencv.org/runk/d7/d8b/tutorial_py_face_
[28] K. W. Bowyer and M. J. Burge, Handbook of Iris Recognition. London, detection.html
U.K.: Springer-Verlag, 2016. [46] Web Link to Download the Source Code and Executable Files for the
[29] H. Proenca and J. C. Neves, “Deep-PRWIS: Periocular recognition Approach Detailed in This Paper. Accessed: Jul. 2020. [Online]. Avail-
without the iris and sclera using deep learning frameworks,” IEEE Trans. able: http://www.comp.polyu.edu.hk/%7ecsajaykr/irisperifusion.htm
Inf. Forensics Security, vol. 13, no. 4, pp. 888–896, Apr. 2018. [47] G. W. Quinn, G. W. Quinn, P. Grother, and J. Matey, IREX IX Part One:
[30] Q. Zhang, H. Li, Z. Sun, and T. Tan, “Deep feature fusion for iris and Performance of Iris Recognition Algorithms. Gaithersburg, MD, USA:
periocular biometrics on mobile devices,” IEEE Trans. Inf. Forensics NIST, 2018.
Security, vol. 13, no. 11, pp. 2897–2912, Nov. 2018. [48] A. Kumar, V. Kanhangad, and D. Zhang, “A new framework for adap-
[31] P. A. Johnson, P. Lopez-Meyer, N. Sazonova, F. Hua, and S. Schuckers, tive multimodal biometrics management,” IEEE Trans. Inf. Forensics
“Quality in face and iris research ensemble (Q-FIRE),” in Proc. 4th IEEE Security, vol. 5, no. 1, pp. 92–102, Mar. 2010.
Int. Conf. Biometrics Theory, Appl. Syst. (BTAS), Sep. 2010, pp. 1–6.
[32] CASIA.v4 Database. Biometrics Ideal Test. Accessed: Oct. 2012.
[Online]. Available: http://www.idealtest.org/dbDetailForUser.do?id=4
[33] D. L. Woodard, S. Pundlik, P. Miller, R. Jillela, and A. Ross, “On the Kuo Wang received the B.E. degree in electronic
fusion of periocular and iris biometrics in non-ideal imagery,” in Proc. engineering and the M.Sc. degree in software engi-
20th Int. Conf. Pattern Recognit., Aug. 2010, pp. 201–204. neering from The Hong Kong Polytechnic University
[34] K. B. Raja, R. Raghavendra, M. Stokkenes, and C. Busch, “Multi-modal in 2013 and 2015, respectively, where he is currently
authentication system for smartphones using face, iris and periocular,” pursuing the Ph.D. degree with the Department of
in Proc. Int. Conf. Biometrics (ICB), May 2015, pp. 143–150. Computing. His research interests include biomet-
[35] G. Santos, E. Grancho, M. V. Bernardo, and P. T. Fiadeiro, “Fusing rics, computer vision, pattern recognition, and deep
iris and periocular information for cross-sensor recognition,” Pattern learning.
Recognit. Lett., vol. 57, pp. 52–59, May 2015.
[36] S. Verma, P. Mittal, M. Vatsa, and R. Singh, “At-a-distance person
recognition via combining ocular features,” in Proc. IEEE Int. Conf.
Image Process. (ICIP), Sep. 2016, pp. 3131–3135.
[37] K. Ahuja, R. Islam, F. A. Barbhuiya, and K. Dey, “A preliminary
study of CNNs for iris and periocular verification in the visible spec- Ajay Kumar (Fellow, IEEE) received the Ph.D.
trum,” in Proc. 23rd Int. Conf. Pattern Recognit. (ICPR), Dec. 2016, degree from The University of Hong Kong,
pp. 181–186. Hong Kong, in 2001. He was an Assistant Profes-
[38] V. Talreja, M. C. Valenti, and N. M. Nasrabadi, “Multibiometric secure sor with the Department of Electrical Engineering,
system based on deep learning,” in Proc. IEEE Global Conf. Signal Inf. IIT Delhi, Delhi, India, from 2005 to 2007. He is
Process. (GlobalSIP), Nov. 2017, pp. 298–302. currently working as a Professor with the Depart-
[39] A. Joshi, A. K. Gangwar, and Z. Saquib, “Person recognition based on ment of Computing, The Hong Kong Polytechnic
fusion of iris and periocular biometrics,” in Proc. 12th Int. Conf. Hybrid University. His current research interests include
Intell. Syst. (HIS), Dec. 2012, pp. 57–62. biometrics with an emphasis on hand biometrics,
[40] J. Cambier, Biometric Data Interchange Formats—Part 6: Iris Image vascular biometrics, iris, and multi-modal biomet-
Data, Standard ISO/IEC 19794, 2011. rics. He holds seven U.S. patents, and has authored a
[41] V. Mnih, N. Heess, and A. Graves, “Recurrent models of visual book on Contactless 3D Fingerprint Identification. He has served on the IEEE
attention,” in Proc. Adv. Neural Inf. Process. Syst., 2014, pp. 2204–2212. Biometrics Council as the Vice President for publications from 2011 to 2015.
[42] P. J. Phillips et al., “FRVT 2006 and ICE 2006 large-scale experimental He is currently serves on the Editorial Board of the IEEE T RANSACTIONS ON
results,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 5, PATTERN A NALYSIS AND M ACHINE I NTELLIGENCE and was on Editorial
pp. 831–846, May 2010. Board of the IEEE T RANSACTIONS ON I NFORMATION F ORENSICS AND
[43] K. H. M. Cheng and A. Kumar, “Advancing surface feature encoding S ECURITY from 2010 to 2013. He has also served on the program committees
and matching for more accurate 3D biometric recognition,” in Proc. 24th of several international conferences and workshops in the field of his research
Int. Conf. Pattern Recognit. (ICPR), Aug. 2018, pp. 3501–3506. interest.
Authorized licensed use limited to: BMS College of Engineering. Downloaded on November 27,2020 at 08:08:09 UTC from IEEE Xplore. Restrictions apply.