This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
1
Biometric Palmprint Verification: A
Dynamical System Approach
David Palma, Pier Luca Montessoro , Giulia Giordano, and Franco Blanchini
Abstract—Most of the existing techniques for palmprint recognition rely on metrics, typically based on static functions, which
evaluate the distance between a pair of features. In this paper,
we propose a new technique for palmprint verification based on
a dynamical system approach for principal palm lines matching.
The proposed dynamic algorithm is recursive and involves a positive linear dynamical system, whose evolution depends on the
matching level between the two input images. In a preprocessing
phase, the procedure iteratively erodes both of the images to be
compared, by eliminating points in each image that do not have
enough close neighboring points both in the image itself and the
comparison image. As a result of the iterations, only the points
that have enough neighboring points in both the image itself and
in the comparison image can survive. Thus, the output of the
dynamical system converges either to zero, when a deep mismatch exists between the two images, or to a high value, when a
good matching is observed. The results, in terms of verification,
are in line with the state-of-the-art results in the current literature. The main advantage of the approach is its robustness when
dealing with low-resolution and noisy images. The impact of noise
(e.g., salt and pepper noise) is effectively reduced: images corrupted with such noise are easily recognized, while a randomly
generated image is rejected even when compared with itself.
Index Terms—Biometrics, dynamic algorithm, dynamical system, line matching, palmprint, principal lines.
I. I NTRODUCTION
ITH the increasing demand of automated systems
based on person identification, the importance of biometric systems is growing. These systems are based on
the measurable biological (anatomical and physiological) or
behavioral characteristics used for the identification of an individual. Different features are used in biometric systems, such
as fingerprints [20], [23], palmprint [4], [7]–[9], [12]–[14],
[16], [17], [19], [21], [24], [25], [28]–[32], [35]–[42], [44],
[46], [47], [52]–[63], hand geometry [26], [31], [45], [49],
iris [1], [10], [51], and face [11], [18], [33], [48], [50]. Unlike
W
Manuscript received January 16, 2017; revised May 2, 2017; accepted
October 12, 2017. This paper was recommended by Associate Editor
D. Zhang. (Corresponding author: Pier Luca Montessoro.)
D. Palma and P. L. Montessoro are with the Polytechnic Department
of Engineering and Architecture, University of Udine, 33100 Udine, Italy
(e-mail:
[email protected];
[email protected]).
G. Giordano is with the Delft Center for Systems and Control,
Delft University of Technology, 2628 Delft, The Netherlands (e-mail:
[email protected]).
F. Blanchini is with the Department of Mathematics, Computer
Science and Physics, University of Udine, 33100 Udine, Italy (e-mail:
[email protected]).
Color versions of one or more of the figures in this paper are available
online at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/TSMC.2017.2771232
conventional methods for personal identification, such as personal identification number, password, and key, these features
cannot be duplicated, lost or stolen. Hence, biometric systems
are suitable to be used in various fields, such as high security,
forensic, and commercial applications.
Palmprint recognition, a relatively novel but promising biometric technology, has recently received considerable interest,
mostly for its importance in forensics [7] (about 30% of the
latents found in crime scenes are from palms [22]) and for several potential civil applications [27]. Compared with the other
physical biometric characteristics, palmprint authentication has
several advantages: low-resolution imaging, low-intrusiveness,
stable line features, and low-cost capturing device. Indeed,
since the principal lines and wrinkles can be observed under
low-resolution images (e.g., 100 dpi or lower), palmprint systems do not require high resolution capturing devices [42].
However, palmprint recognition techniques based on both lowand high-resolution features are proposed in the literature (see
the survey in Section II); such methods use as local features, respectively, principal lines and ridges. The approach
discussed in this paper is based on low-resolution features,
since it uses as local features the principal lines of the palm,
which are very important physiological characteristics to distinguish between different individuals because of their stability
and uniqueness. Compared with ridges and wrinkles, principal
lines are usually the consequence of genetic effects: therefore, they are the most significant features in palmprint images
and have good permanence. However, principal lines may be
similar in different individuals, which makes their distinctiveness relatively low; for this reason, palmprint recognition is a
challenging problem [60].
To address this challenge, we propose a method consisting
of the following.
1) A region of interest (ROI) extraction phase, which follows the typical sequence of steps used in [35] and [60],
to face different issues mainly due to nonlinear distortion, such as rotation and translation of the palm with
respect to the image and nonuniform illumination.
2) An unconventional feature extraction phase based on the
principal lines of the palmprint [63].
3) A novel approach to palmprint matching based on a
dynamic algorithm. The algorithm involves a positive
linear dynamical system [3], whose evolution is determined by the matching level between the two input
images: its output converges to zero when the two
images have a deep mismatch, while it reaches a high
value in the case of good matching.
c 2017 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
2168-2216
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
2
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
Methods based on dynamical systems have been successfully
exploited to improve the performance of algorithms (see [2]).
Here, we show that the proposed dynamic algorithm, included
as a step in the matching phase, confers more robustness
to palmprint recognition techniques, especially when lowresolution and noisy images are involved: the impact of noise,
such as salt and pepper noise, is effectively reduced, avoiding
both false rejection of images corrupted by noise and false
acceptance of randomly generated images.
The results obtained by means of this novel algorithm are
promising, and comparable to existing palmprint recognition
algorithms. Furthermore, in order to allow implementation on
mobile devices for online user authentication, each step has
been specifically designed to limit memory requirement and
computational complexity.
Section II of this paper provides an overview of the related
work. Then, Section III presents the new approach to palmprint
verification based on a noise-rejecting dynamic algorithm,
while Section IV describes the main steps of the preprocessing
phase and feature extraction phase that is based on the principal lines of the palmprint. Section V reports and discusses the
experimental results of the developed palmprint verification
system. Section VI draws conclusions.
II. R ELATED W ORK
In recent years, palmprints have been investigated extensively in automated personal authentication, because of their
advantages with respect to other physical biometric characteristics. Therefore, a number of interesting palmprint recognition
techniques have been proposed in the literature, which can
be grouped in two categories: 1) approaches based on lowresolution features and 2) approaches based on high-resolution
features, which use creases and ridges, respectively, as local
features. The former approach analyzes the structure of the
palmprint, where generally only principal lines, wrinkles, and
texture are well evident, while the latter approach tries to find
the ridges or minutiae-like point features to represent the palm
pattern. Palm lines include both principal lines and wrinkles,
but principal lines get more attention because of their stable
and highly distinctive nature [5], [34]. Hence, most authors
focus on palm lines when analyzing palmprint images, and
we will also focus on low-resolution methods in the following.
In line-based approaches, the extracted palm lines are either
matched directly or represented in other formats through some
feature transformations for matching.
Zhang and Shu [62] used the datum point invariant property and the line feature matching technique to conduct the
verification process via the palmprint features. This approach,
in which the authors inked the palmprints on to paper and
then scanned them to obtain images, is not suitable for many
online security systems. Wu et al. [54] used the Canny operator [6] to detect palm lines and the orientations of the edge
points are passed into four membership functions representing
four directions. For each direction the energy is computed,
then the feature value is normalized. Finally, Euclidean distance is used for matching. Wu et al. [57] also proposed
an algorithm based on the hidden Markov models (HMMs).
This approach uses Sobel masks to compute the magnitude
of palm lines, then the projections along both x and y directions are used to obtain the inputs of the HMMs. A third
approach proposed by Wu et al. [55] is based on the design of
two masks which compute the vertical first- and second-order
derivatives obtained by rotating the two standard masks. They
identify the edge points and corresponding directions by using
the zero-crossing of the first-order derivative, then the positive magnitude of the corresponding second-order derivative
is considered as the magnitude of the lines. The feature vector,
consisting of the weighted sum of the local directional magnitude, is normalized by its maximum and minimum elements.
Finally, as in [54], Euclidean distance is used for matching.
Boles and Chu [4] used Sobel masks and thresholds to construct binary edge images and then the Hough transform to
detect the palm features as approximated straight lines for
matching. Huang et al. [21] and Jia et al. [25] used the modified finite Radon transform (MFRAT) to extract the principal
lines, then a matching algorithm based on pixel-to-area comparison measures the similarity between two palmprint images.
Fei et al. [16] used a double half-orientation-based method
for feature extraction and palmprint recognition, while in [17]
they propose a double-orientation code (DOC) algorithm to
represent the orientation of palmprint features, evaluating
the similarity between the DOC through a nonlinear angular
matching scheme. Gao and Leung [18] and Leung et al. [36]
made use of Sobel masks to extract line-like features from
the palmprints and then of line segment Hausdorff distance
to compare two palmprints. Han et al. [19] employed Sobel
and morphological operations to extract line-like features from
palmprint images obtained using a scanner, and then a correlation function to measure the similarity between the two
feature vectors. Similarly, for verification, Kumar et al. [31]
used other directional masks to extract line-like features from
the palmprints captured using a digital camera, then use a
normalized correlation for matching. Diaz et al. [12] used
Sobel masks and morphologic operator as two separated feature extractors to obtain the gradient of the images, which are
then classified by neural networks. Kung et al. [33] designed a
decision-based neural network classifier and applied it to face
recognition and to palmprint verification. Other approaches
based on low-resolution features are named subspace-based
approaches, which usually adopt principal component analysis
(PCA), linear discriminant analysis (LDA), and independent component analysis (ICA). Zhang et al. [60] used 2-D
Gabor filters to extract texture features from low-resolution
palmprint images captured using a CCD camera: two palmprint images are compared in terms of their Hamming distance. Connie et al. [8] made use of PCA and ICA to
extract the palmprint texture features, then compare palmprints based on three types of classifiers: 1) Euclidean distance;
2) cosine measure; and 3) probabilistic neural network methods. Ekinci and Aykut [14] used a wavelet-based kernel PCA
method by integrating the Daubechies wavelet representation
of palm images and the kernel PCA method for palmprint
recognition; similarity measurement is accomplished by using
a weighted Euclidean linear distance-based nearest neighbor
classifier.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
PALMA et al.: BIOMETRIC PALMPRINT VERIFICATION: DYNAMICAL SYSTEM APPROACH
3
Apart from neural networks, to the best of the authors’
knowledge, dynamic algorithms have never been employed for
palmprint matching. The novel method we propose is based
on the evolution of a suitable dynamical system that erodes
nonmatching features and enhances matching features.
III. DYNAMIC A LGORITHM FOR M ATCHING
This section highlights the main idea of the proposed noiserejecting dynamic (iterative) algorithm for palmprint matching.
Given two images (matrices of Boolean values) X, Y ∈
{0, 1}s×s , of equal dimensions, we consider the operator
′ ′
(1)
X , Y = f (X, Y)
which provides two new images X ′ , Y ′ ∈ {0, 1}s×s , having the
same size as X and Y. For our palmprint verification purposes,
the operator f must be chosen to assess the “matching level”
between the two initial images. To simply explain the idea,
let us consider a simple (non dynamic) mismatch function.
Denoting by (·) the number of 1 (active) pixels in an image,
we define the matching index between X and Y as
Y′
γ (X, Y) X ′
+
(2)
α(X, Y) =
2
(X)
(Y)
where the coefficient γ is lower when the difference of active
pixels in the input images is higher
⎧
(X)
⎪
⎪
if (X) ≥ (Y)
⎨1 − logs
(Y)
γ (X, Y) =
(3)
(X)
⎪
⎪
1
+
log
otherwise.
⎩
s
(Y)
With this choice of γ , for large differences in the number of
active pixels between the two initial images, the coefficient
quickly tends to a low value, thus reducing α(X, Y).
To assess the matching level of two images, a simple option
is to consider the intersection X ′ = Y ′ = X∩Y, associated with
the AND logic function
X
Y = xij AND yij .
ij
However, the approach based on this choice of (1) is clearly
not noise-rejecting, since it evaluates the correspondence
between single pixels (isolated from their context), which may
be noisy. Conversely, to avoid a misleading recognition, noisy
pixels that are accidentally matching in the two images should
not provide a positive contribution to the matching index.
We pursue an approach that rejects noise and, to this aim,
we assume that the matching of “isolated spots” is not as significant as the matching of wide “stripes” or “islands,” even
if the number of matching isolated spots is very high. The
operator f must then be chosen so as to set to 0 all the pixels
without a sufficient number of neighboring 1 pixels, both in
the image itself and in the comparison image: only significant clusters of pixels that have a corresponding cluster in the
complementary image must remain active.
The desired behavior is illustrated in Fig. 1.
1) In the first example, two significant sets of active pixels
are surrounded by 0 pixels in both images X and Y,
which show a good matching. Actually, almost all of the
Fig. 1. Desired matching patterns: 1 good matching; 2 partial matching;
and 3 no matching. X and Y are the initial images, Z = X ′ and W = Y ′ the
desired transformed images.
active pixels should survive in the transformed images
Z = X ′ and W = Y ′ .
2) In the second example, there is a just partial superposition. Hence, in both of the transformed images, the sets
of active pixels should be partially eroded.
3) In the third example, there is a perfect superposition
between isolated points. In this case, the transformed
images should be almost empty.
As is experimentally shown later, the behavior in Fig. 1 can
be achieved with the proposed dynamic (iterative) algorithm.
Remark 1: In view of the iterative nature of the algorithm,
we only need to emphasize the presence of neighborhoods in
the comparison image, since in at most two steps this indirectly
emphasizes neighborhoods in the image itself.
We assume that the images A and B, provided as inputs,
have already undergone a preliminary processing (thoroughly
described in Section IV). Given the set
N ij = {h, l : |h − i| ≤ ν, |l − j| ≤ ν, h, l ∈ Z}
of neighboring points of (i, j) within a “radius” of integer
amplitude ν > 0, the “fading factor” 0 < λ < 1, the coefficient µ > 0 that emphasizes the presence of neighbors in
the comparison image and the tolerance ǫ > 0, Algorithm 1
works as follows.
Note that the images are processed as real-valued matrices,
in view of step 1. Due to the componentwise product at step 4,
all the pixels that are initially 0 remain 0: the procedure can
“inactivate” active pixels, but cannot “activate” inactive pixels.
As a consequence, the index α is always a number 0 ≤ α ≤ 1.
The rationale of step 3 (the core of the procedure) is the following. Assume that a pixel has no match on the other image.
Then the corresponding equation becomes
xij (k + 1) = λxij (k).
Hence, being λ < 1, the value of pixels having no match on
the other image quickly converges to zero: typically, in few
iterations xij (k) goes below the threshold ǫ. We stress that, if
a pixel has no match at the beginning, this status is preserved
for all the steps of the procedure, due to step 4.
Conversely, the value of pixels having a large neighborhood in the comparison image diverges. If the same condition
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
4
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
Algorithm 1 Matching Index Computation
Input: Boolean images A and B.
Parameters: Number of steps N, positive constants λ < 1 and
µ, integer neighborhood amplitude ν > 0, a small tolerance
ǫ > 0.
Outputs: Matching index α(A, B).
1) Convert the two input images from boolean into real
matrices X := A and Y := B.
2) Set k = 0.
3) At each iteration, compute the updated values for each
pixel in both images
yhl (k)
(4)
xij (k + 1) = λxij (k) + µ
(a)
hl∈Nij
yij (k + 1) = λyij (k) + µ
xhl (k)
(5)
hl∈Nij
4) At each iteration, set X := X × A, Y := Y × B, where ×
denotes the componentwise product (i.e., xij := xij × aij
for all i, j).
5) Set k = k + 1 and, IF k < N, GOTO step 3.
6) Convert the matrices to boolean, [A′ , B′ ] = logic(X, Y),
as follows: IF xij < ǫ, THEN a′ij := 0; ELSE a′ij := 1
(resp. IF yij < ǫ, THEN b′ij := 0; ELSE b′ij := 1).
7) Compute the matching index α as in (2).
(a)
(b)
(c)
Fig. 2. Dynamic algorithm behavior. (a) and (b) Similar principal lines
extracted from two different subjects, provided as input, and (c) image
resulting from the application of the dynamic algorithm.
is true in the complementary neighborhood, then we are in
the presence of two matching groups that “cooperate” by
increasing each other. In this case, the values of xij and yij
significantly increase in few iterations. The exponential convergence/divergence of the procedure assures that, in few
iterations, some pixels become practically zero (eventually set
to 0), and others achieve large values (eventually set to 1).
To exemplify the algorithm behavior, suppose that the two
input images A and B are those shown in Fig. 2(a) and (b).
Then, the result of the dynamic algorithm, A′ = B′ , is shown in
Fig. 2(c). In both images, the algorithm has iteratively eliminated points that do not have enough close neighboring points
in the comparison image; the number of survived pixels is
very limited, even when starting from images having similar
principal lines.
A considerable advantage of our dynamic approach is its
robustness with respect to noise. Images corrupted, e.g., with
salt and pepper noise are easily recognized, while an image
randomly generated is rejected even if compared with itself.
(b)
Fig. 3. Salt-and-pepper noise. (a) A and B (A = B). (b) A′ and B′ (A′ = B′ ).
Fig. 4. Dynamic algorithm example: (top) pairs of isolated points and (bottom) clusters of points. Points in the same positions are connected through
solid lines, whereas points in different positions but close enough (the radius
is ν = 1) are connected through dashed lines.
Example 1 (Salt-and-Pepper Noise): To explain a peculiar
feature of the algorithm, we compare a salt-and-pepper noisy
image with itself (see Fig. 3). The algorithm eliminates almost
all of the pixels and the resulting matching index is extremely
low.
The idea of the procedure can be described as follows.
Consider a family of bipartite graphs, whose nodes are the
active pixels in the two images (A and B) to be compared. An
arc connects each pair of active pixels that occupy neighboring
positions in the two images. In this way, we form clusters of
active pixels. A cluster is preserved only if it includes enough
pixels, otherwise it fades away. In the case of perfect matching
of two images with isolated points (as in Fig. 4, top), there
are just isolated pairs of nodes, which disappear when the
dynamic algorithm evolves1 (local stability and convergence
to 0). Conversely, in the presence of partially matching blobs
1 Although the term “to evolve” is used, the algorithm is not based on
machine-learning models. We want to emphasize the iterative nature of the
dynamic algorithm that converges to the final state by eroding the pixels that
do not match within a neighborhood.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
PALMA et al.: BIOMETRIC PALMPRINT VERIFICATION: DYNAMICAL SYSTEM APPROACH
5
(as in Fig. 4, bottom), the clusters of nodes are numerous
enough to survive (local instability with saturation to 1).
A. Properties of the Algorithm and Parameter Tuning
We analyze here the algorithm based on the linear recursive
equations (4) and (5), to better understand its behavior.
First of all, the parameters λ and µ in the procedure need
to be tuned. We remind that:
1) the fading parameter λ < 1 cancels “isolated pixels”;
2) the emphasis parameter µ increases “surrounded pixels.”
Consider a generic neighborhood N of size ν and assume that
all the points in N are initially active (set to 1) in both A and
B. Then, their values should “blow up.” Given the updating
equation for xij
xij (k + 1) = λxij (k) + µ
yhl (k)
Fig. 5.
Two sets of pixels with separate evolutions.
Note that the matrix
hl∈Nij
let m be the number of pixels in the considered neighborhood.
Being all the pixels initially set to 1, a blow-up condition is
λ + µm > 1
(6)
(the lower bound in Fig. 9). Therefore, at step k = 1, all the
surrounded pixels reach a value greater than 1; at step k = 2,
the value becomes even bigger; an so on.
Let us now consider a partial superposition, in which only
a percentage p < 1 of the pixels in the neighborhood are
1. We may take p as a threshold fraction, so that 1-pixels
with less than a fraction p of 1-pixels in the complementary
neighborhood tend to fade. This leads to the criterion
λ + pµm < 1
(7)
(the upper bound in Fig. 9). Therefore, given the neighborhood
radius ν and the number of iterations N, a criterion to choose
the parameters λ and µ is the following.
1) Fix λ based on the desired fading speed.
2) Compute m, the neighborhood size, in terms of pixels.
3) Fix p, the desired threshold fraction.
4) Compute µ which satisfies (6) and (7).
Property 1 (Decoupling): Different regions that are not connected in either A or B do not interact and have separate
evolutions, as illustrated in Fig. 5. Hence, isolated spots can
be analyzed separately.
Property 2 (Convergence/Divergence): Two overlapping
spots in A (containing q pixels) and in B (containing s pixels)
give rise to a positive linear system, as described next. Let us
group in a vector x(k) ∈ Rq the variables xij of the spot in A
and in a vector y(k) ∈ Rs the variables yij of the spot in B.
The system evolves as follows:
λI
µ x(k)
x(k + 1)
=
(8)
y(k + 1)
µ⊤ λI y(k)
where I is the identity matrix (of the proper dimension), λ and
µ are the previously fixed constants and is a connection
matrix defined as
1, if pixel yj is in the neighborhood of pixel xi
ij =
0, otherwise.
λI
M=
µ⊤
µ
λI
appearing in (8) is symmetric. In fact, if pixel yj is in the
neighborhood of pixel xi , also xi is in the neighborhood of yj .
Then, the fading or the permanence of the two connected
spots is doomed by the convergence or divergence of the linear system (8). The linear system converges (i.e., is stable)
if and only if the dominant eigenvalue σmax (i.e., the eigenvalue with maximum modulus) of matrix M is smaller than 1.
Conversely, the linear system diverges (i.e., is unstable) if and
only if σmax is greater than 1. Since matrix M is non-negative,
its dominant eigenvalue σmax is real and positive. Moreover,
σmax is an increasing function of both λ and µ, and increases
if the number of 1 entries in matrix increases.
This explains why the pair of identical images with salt and
pepper noise leads to a poor matching index: even if there is
actually a good (or perfect) matching between the two images,
there are not enough connections between the active pixels of
one image and the active pixels of the other.
Since in the average, for the salt and pepper noise, half of
the pixels in the complementary neighborhood are inactive,
we recommend to choose
p ≈ 1/2.
Then, in view of (6) and (7), this leads to the conditions
1−λ
1
<
< 1.
2
µm
IV. PALMPRINT I MAGE P ROCESSING
Fig. 6 shows the overall block diagram of palmprint image
processing that consists of two main phases: 1) preprocessing
and 2) feature extraction.
A. Preprocessing
Before the feature extraction step, a preprocessing is essential in order to obtain a subimage denominated ROI from the
captured palmprint image. In fact, usually palmprint images
can have different orientation and size, and are also subject
to noise. Moreover, the region of not-interest (e.g., fingers,
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
6
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
(a)
Fig. 6.
(b)
(c)
Block diagram of palmprint image processing.
wrist, and image background) may affect the accuracy in processing and verification performance. First of all, noise is
reduced by proper filters, then the ROI is extracted. In order
to align the different palmprint images, 1) binarization and
2) edge detection are executed. After alignment 3) ROI is
extracted and 4) normalized to have a specific mean and variance for all images [20]. These four steps are shown in Fig. 7
and described more in detail below.
1) Binarization: Binarization converts a gray level image
I(x, y) into a binary image B(x, y), where the hand is the foreground and the rest is background. Thus, a local adaptive
thresholding can be applied to binarize the image.
After partitioning the image into nonoverlapping sub-blocks
of size N×N, for each sub-block we compute a threshold value
by using Otsu’s method [43]: the image pixels are split into
two classes (background and object), and the best threshold
value t is obtained based on the maximum value of the variance
between the two classes [15]. Then, for each sub-block, we
apply a global thresholding as follows:
1 if I(x, y) ≥ t
B(x, y) =
(9)
0 otherwise.
The image after the binarization phase is shown in Fig. 7(b).
2) Edge Detection: To detect the hand edge, the segmented
binary image is filtered by using the Canny operator [6]. This
edge detector ensures good noise immunity and at the same
time detects true edge points with minimum error [15]. A
practical approximation of the Canny filter is the following.
1) Convolve the input image I(x, y) with a 2-D Gaussian
filter G(x, y)
H(x, y) = I ∗ G(x, y).
(10)
In order to reduce the computational cost, a convolution with a 2-D Gaussian can be separated into two
convolutions with 1-D Gaussians.
2) Apply the gradient operator and calculate the gradient
magnitude M and orientation
as follows:
∂H(x, y)
∂H(x, y)
Hy (x, y) =
∂x
∂y
M(x, y) = Hx2 (x, y) + Hy2 (x, y)
Hx (x, y) =
(x, y) = arctan
My (x, y)
.
Mx (x, y)
(11)
(12)
(13)
3) Apply the nonmaximum suppression algorithm that
removes the nonmaximum pixels, while preserving the
connectivity of the contours, as follows: for each pixel
(d)
(e)
(f )
Fig. 7. Main steps of preprocessing. (a) Original palmprint image. (b) Noise
reduction and binarization by using local adaptive thresholding. (c) Edge
detection. (d) Scaling, alignment and ROI location. (e) ROI extraction.
(f) Normalization.
(x, y), if one of the neighboring pixels in the orthogonal
directions to edge orientation is greater than the current
pixel, then discard it.
4) The last step is the hysteresis thresholding, which uses
two thresholds: TH , calculated by using Otsu’s algorithm [15], and TL = (1/2)TH . This algorithm sets a
pixel (x, y) to 0 if it has a value less than TL , and to 1
if it has a value greater than TH or it is connected to a
pixel that has a value greater than TH .
The image after the edge detection phase is in Fig. 7(c).
3) Region of Interest Extraction: In this last step, the palmprint image is aligned in a standard pose, to reduce issues due
to nonlinear factors, such as nonuniform illumination, rotation, and translation. In order to extract the central part of
the palmprint images, we detect two reference points between
the fingers, as shown in Fig. 7(d). These points are used to
construct a reference line for aligning the different palmprint
images, and the middle point between them is used to determine the central position of the ROI. Finally, the ROI can be
extracted as shown in Fig. 7(e).
4) Normalization: In order to have a prespecified mean and
variance for all palmprint images, the extracted ROI images
are normalized to reduce the possible nonuniform illumination
and noise, as suggested in [23]. Therefore, given the input ROI
image I(x, y), the normalized image is defined as follows:
µn + ρ if I(x, y) > σ 2
′
(14)
I (x, y) =
µn − ρ if I(x, y) ≤ σ 2
where
ρ=
σn2 (I(x, y) − µ)2
.
σ2
(15)
The parameter values µn and σn2 used for the computation
are the same suggested in [23], thus we set both the mean and
variance values to 100. The result is shown in Fig. 7(f).
B. Feature Extraction
Once the central part (ROI) is segmented, features can be
extracted for matching. The feature extraction phase plays an
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
PALMA et al.: BIOMETRIC PALMPRINT VERIFICATION: DYNAMICAL SYSTEM APPROACH
important role in image identification and verification. There
are many features exhibited in a palm, such as the three principal lines caused by flexing hand and wrist in the palm, which
are denominated as heart line, head line, and life line, respectively. We use a principal line extraction method based on [63],
whose main steps are illustrated in Fig. 6: 1) conversion to
a negative image; 2) sharpening of the image with Top-Hat
filter and gradient magnitude computation; 3) linear contrast
adjustment; 4) binarization with Otsu’s method; and 5) noise
cleaning with median filter.
1) Negative: After normalization in the last step of the preprocessing phase, the resulting enhanced ROI image I(x, y) is
converted to its negative as follows:
I ′ (x, y) = max{I(x, y)} − I(x, y).
(16)
In this way, the principal lines of the palm have high intensity.
2) Sharpening: To correct uneven illumination we use a
Top-Hat transform, then apply a linear contrast enhancement to
the output image. The Top-Hat transform enables us to extract
small bright objects from a varying background and is defined
as the difference between the input image and its morphological opening by using a defined structuring element ρ. The
opening of I(x, y) by ρ is obtained by the erosion of I(x, y)
by ρ, followed by dilation of the resulting image by ρ:
I ◦ ρ = (I ⊖ ρ) ⊕ ρ =
I(x, y).
(17)
7
(a)
(b)
Fig. 8. Some results of palm line extraction. (a) Feature extraction on CASIA
database. (b) Feature extraction on PolyU database (lower resolution).
5) Noise Cleaning: As a last step in the feature extraction
phase, a median filter is used in order to remove noise and trivial lines from the image. Fig. 8 shows examples of extracted
feature with the algorithm described above.
I(x,y)⊂ρ
Hence, the opening of I(x, y) by ρ is the union of all sets
I(x, y) contained in ρ. Then, the opening Top-Hat transformation of image I, denoted by TH, is given by
TH = I − (I ◦ ρ)
(18)
that represents the difference set of the domain of I(x, y) and
the opening domain of I(x, y). After the morphological operations, the gradient magnitude is computed as described in
the preprocessing phase: this sharpening operation returns the
bright spots of the image that are smaller than the structuring
element.
3) Contrast Adjustment: In order to improve the contrast of
the image, a linear contrast enhancement is applied by stretching the range of intensity values it contains to a specific range
of values. Before applying the linear scaling function it is
necessary to identify the lower and upper bounds from the
histogram, which are the minimum and maximum brightness
values in the image, and to apply a transformation to stretch
this range to fill the full range (0, 255). Let min and max be
the lower and the upper limits of the histogram, respectively,
then each pixel of the image is scaled to the range (0, 255)
through the following equation:
I ′ (x, y) = (I(x, y) − min)
255
.
max − min
(19)
4) Binarization: A global thresholding is applied at the
gray level image resulting from the previous sharpening filter and linear contrast adjustment. The threshold is calculated
with Otsu’s method [43] described in the preprocessing phase.
V. E XPERIMENTAL R ESULTS
A. Databases Used in Simulation
The performance of the proposed palmprint verification
system has been tested upon two databases that have been
worldwide shared for research purposes: 1) CASIA palmprint
database [65] and 2) PolyU palmprint database II [64].
The first database contains 5502 palmprint images with 8
bit gray-level of size 640 × 480 pixels at 96 dpi resolution
captured from 312 subjects by a low resolution CCD-based
device. For each subject there are palmprint images from both
left and right hand captured eight times at different times from
people of different ages. The second database, which has a
lower resolution, contains 7752 palmprint images with 8 bit
gray-level of size 384×284 pixels at 75 dpi resolution captured
from 386 palms of 193 subjects: these palmprint images have
been collected in two sessions with the average interval over
two months, where about ten samples have been captured from
each palm for each session.
We have performed experiments by taking: 1) 16 sample
palmprint images of the left and right hands (eight images for
hand) of all 312 subjects, for a total of 4992 images, from the
CASIA palmprint database and 2) ten sample palmprint images
of 381 individuals related to the first and second sessions, for
a total of 7620 images, from the PolyU palmprint database II.
This setup
constitutes
a total number of experiments
equal
16
=
37
440
=
12
457
536,
including
312
×
to: 1) 4 992
2
2
genuine experiments (the others are impostor
experiments),
= 29 028 390,
for the CASIA palmprint database and 2) 7 620
2
20
including 381 × 2 = 72 390 genuine experiments, for the
PolyU palmprint database.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
8
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
B. Criteria for Performance Evaluation
To evaluate the verification accuracy of the proposed
method, each palmprint image in the database has been
matched against the other palmprint images. The matching
between palmprints captured from the same palm is defined
as a genuine matching or intraclass matching, otherwise as an
impostor matching or interclass matching.
The general method to evaluate the performance of a palmprint authentication system is based on the false acceptance
rate (FAR) and the false rejection rate (FRR) [41]. The FAR
is defined as the percentage of invalid inputs which are incorrectly accepted and is computed as the number of accepted
imposter claims over the total number of imposter accesses.
The FRR is defined as the percentage of valid inputs which are
incorrectly rejected and is computed as the number of rejected
genuine claims over the total number of genuine accesses. FAR
and FRR are closely related because the increase of the one
implies the decrease of the other. Mathematically these two
indices are expressed as follows:
FAR =
FRR =
where
N
1
FAR(k)
N
(20)
1
N
(21)
k=1
N
FRR(k)
Fig. 9. Region of convergence of the system delimited by 6 and 7 [with
m = 25 and p = (1/2)], and values of the parameter pairs (λ, µ). Best results
have been obtained by the highlighted pair of values.
k=1
number of FARs
number of impostor accesses
number of FRRs
FRR(k) =
.
number of genuine accesses
FAR(k) =
(22)
(23)
Other evaluation parameters are as follows.
1) Receiver Operating Characteristic (ROC): A graphical
plot that illustrates the tradeoff between the FAR and
the FRR.
2) Equal Error Rate (EER): The error rate at the specific threshold s for which both acceptance and rejection
errors are equal, which can be easily obtained from the
ROC curve.
3) Genuine Acceptance Rate (GAR):
GAR = 1 − FRR.
4) GAR at a Specific FAR (GARx ):
GARx = GARFAR=10−x
Fig. 10. Matching value against number of iterations for two palmprint
features in the case of true positive recognition and true negative recognition.
(24)
(25)
for example,
considered an FAR = 0.01, then GAR2 =
GARFAR .
C. Parameter Tuning for Performance Evaluation
Since the verification rate of the system varies depending on
the values of λ and µ in the dynamic algorithm, these parameters need to be properly set, as discussed in Section III. Thus,
given the set Ni,j of neighboring points of the generic point
p(i, j) we can choose the fading factor 0 < λ < 1 and the coefficient µ > 0. Fig. 9 depicts the region of convergence of the
system according to (6) and (7), which represent the upper and
lower bounds of the region, respectively. Since the thickness
of the palm lines consists of a few pixels, it is reasonable to
consider a square matrix with radius ν = 2 (which amounts to
considering a number of neighboring pixels equal to m = 25)
for a perfect coverage of a palm line avoiding excessive overlaps with other palm lines—or background—contained in the
image to be compared.
Therefore, to determine the value of the parameters that
guarantees the best performance, a test over a subset of the
CASIA palmprint database for many different values of the
parameter pairs has been performed. In this phase, we have
verified every test pair for each of the palmprint images in the
database and for each pair of the parameter values.
The subset of the CASIA palmprint database consists of 360
palmprint images and the number of parameter pairs (λ, µ)
inside the region of convergence is equal to 70, thus the
amount of the tests performed is 4 523 400. As a result of
these tests we found that λ = 0.465 and µ = 0.042 provide
the best results. These values have been used in the experiments for both the databases in order to test the performance
of the proposed method, even without parameter retuning for
the PolyU database.
It is also important to choose a fixed number of iterations,
after which we can rely on convergence of the dynamic system.
From Fig. 10 it is clear that, after few iterations, the reached
matching value can be considered good enough with respect
to the highest value achieved. Thus, to save computation time,
the number of iterations for the tests has been fixed to 20.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
PALMA et al.: BIOMETRIC PALMPRINT VERIFICATION: DYNAMICAL SYSTEM APPROACH
9
(a)
(b)
(c)
Fig. 12. Dynamic algorithm behavior on different principal line images from
the same subject. (a) First image affected by white Gaussian noise with µ = 0
and σ 2 = 0.05, (b) second image to be compared, and (c) surviving points
after the algorithm evolution.
Fig. 11. Comparative GAR against FAR graphs, for the proposed approach
and for other methods on PolyU palmprint database.
TABLE II
ROC C OMPARISON ON B OTH THE DATABASES W ITH N ORMAL
PALMPRINTS AND PALMPRINTS C ORRUPTED BY
A DDITIVE W HITE G AUSSIAN N OISE
TABLE I
C OMPARISON OF V ERIFICATION R ESULTS FOR
D IFFERENT M ETHODS [16], [40]
D. Performance Comparison
The performance results of the developed palmprint recognition system are here compared with other algorithms by using
the datasets as previously described.
For comparison purposes, some representative techniques in
palmprint recognition have been used.
1) CompCode [29], which uses six real parts of the Gabor
filters to obtain orientation information from palm lines
and subsequently applies Hamming distance for feature
matching.
2) PalmCode [60], which uses 2-D Gabor phase encoding
scheme to extract the feature, then Hamming distance is
used for matching.
3) OrdiCode [52], that compares two line-like image
regions, which are orthogonal in orientation and it generates one bit feature code, then for matching is also
used the Hamming distance.
4) Huang et al. algorithm [21], which uses the MFRAT
to extract palmprint principal lines, then a matching
algorithm based on pixel-to-area comparison has been
designed to measure the similarity.
5) DOC [17] which stands for double-orientation code used
to represent the orientation feature of palmprint which
are compared with a nonlinear angular matching score
method.
Table I lists the verification results, expressed in terms
of EER, for different approaches, on databases that contain
palmprints captured through touch-based (PolyU) and touchless (CASIA) devices. From the results, it is clear that
the CASIA database is more challenging than the PolyU
database in terms of palmprint verification. The proposed
method achieves good results, outperforming some of the other
approaches.
To better illustrate the verification performances, the ROC
curves of the previous five methods and the proposed algorithm are illustrated in Fig. 11: the ROC curves are obtained
by plotting GAR against FAR. The dynamic algorithm obtains
a GAR2 that is almost 100% and a GAR3 that is greater
than 97%.
E. Robustness to Noise
To conduct experiments on noisy palmprints and demonstrate the robustness of the dynamic algorithm with respect to
noise, the system has been tested by comparing palmprints corrupted by white Gaussian noise against normal palmprints. The
experiments are carried out by corrupting the images with a
white Gaussian noise of mean µ = 0 and variance σ 2 = 0.05.
Fig. 12 depicts an example of the dynamic algorithm behavior:
it can be seen from the output image that the subject is easily
recognized, validating the noise rejection property described
in Section III.
Table II reports the result of noise rejection experiments and
shows that the performance of the system is not that much
degraded: even in presence of noise, the GAR1 is greater than
98.6%. However, we have obtained better results on the PolyU
database achieving a GAR of 97.8% with an EER equal to
0.022 against a GAR of 96.9% with an EER equal to 0.031 on
the CASIA database. Therefore, the experiments show that the
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
10
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
proposed algorithm can recognize palmprints even in presence
of a reasonable amount of noise.
F. Computation Time
The experiments for the proposed approach have been conducted on a virtualized machine equipped with two dedicated
processors and 2048 MB RAM hosted on an Intel Core i74510U CPU (2.6 GHz), 8 GB RAM running a 32-bit Microsoft
Windows 10. The code has been implemented using MATLAB
8.0 and to get the following results each part of code has
been performed for 200 times, then the average time has been
considered. Thus, the computational times required for preprocessing, palm line extraction, and palmprint matching are
376, 49, and 295 ms, respectively, thus the average response
time for verification is about 0.72 s, which is more suitable
for a real-time biometric verification system rather than a realtime biometric identification system. However, it is possible to
further reduce the computation time, since the program codes
have not completely optimized.
VI. C ONCLUSION
In this paper, a novel approach has been presented to authenticate individuals based on their palmprint features. As a main
contribution, a new recursive, dynamic algorithm has been
applied for the matching of features. A noticeable advantage
of such an approach is its robustness with respect to noise:
for instance, images corrupted with salt and pepper noise are
easily recognized, whereas an image randomly generated is
rejected even when compared with itself [44]. The images provided as an input to the dynamic algorithm have undergone an
image processing based on two phases: the first involving preprocessing operations to make the system invariant to rotation
and translation of the palm with respect to the image and the
second consisting of a sequence of robust feature extraction
steps that allow to detect the principal lines of the palm without
introducing noise. Extensive experiments have been conducted
to evaluate the performance of the system and the results
obtained from the tests clearly demonstrate the effectiveness
of the proposed technique, which is able to produce promising
result. As a matter of fact, the experimental results presented
in Section V show that the level of GAR1 obtained is 100% on
both the databases, whereas in the noisy test GAR1 can still
be considered greater than 98.6% for the CASIA database and
equal to 100% for the PolyU database. Thus, the system is
comparable with existing biometric recognition systems based
on palmprint recognition and other hand-based biometric technologies, including hand geometry and fingerprint verification.
In particular, the achieved level of GAR is compliant with the
strict requirements of high security applications.
R EFERENCES
[1] S. Alkassar, W. L. Woo, S. S. Dlay, and J. A. Chambers, “Robust
sclera recognition system with novel sclera segmentation and validation techniques,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 47, no. 3,
pp. 474–486, Mar. 2016.
[2] F. Blanchini, D. de Caneva, P. L. Montessoro, and D. Pierattoni,
“Control-based p-persistent adaptive communication protocol,” ACM
Trans. Auton. Adapt. Syst., vol. 7, no. 2, pp. 1–18, 2012.
[3] F. Blanchini and S. Miani, Set-Theoretic Methods in Control (Systems &
Control: Foundations & Applications). Boston, MA, USA: Birkhäuser,
2008.
[4] W. W. Boles and S. Y. T. Chu, “Personal identification using images
of the human palm,” in Proc. IEEE Reg. 10 Annu. Conf. Speech Image
Technol. Comput. Telecommun., vol. 1. Brisbane, QLD, Australia, 1997,
pp. 295–298.
[5] A. Bruno, P. Carminetti, V. Gentile, M. La Cascia, and E. Mancino,
“Palmprint principal lines extraction,” in Proc. IEEE Biometric Meas.
Syst. Security Med. Appl., Rome, Italy, 2014, pp. 50–56.
[6] J. Canny, “A computational approach to edge detection,” IEEE Trans.
Pattern Anal. Mach. Intell., vol. PAMI-8, no. 6, pp. 679–698, Nov. 1986.
[7] R. Cappelli, M. Ferrara, and D. Maio, “A fast and accurate palmprint
recognition system based on minutiae,” IEEE Trans. Syst., Man, Cybern.
B, Cybern., vol. 42, no. 3, pp. 956–962, Jun. 2012.
[8] T. Connie, A. Teoh, M. Goh, and D. Ngo, “Palmprint recognition with
PCA and ICA,” in Proc. Image Vis. Comput., 2003, pp. 227–232.
[9] T. Connie, A. T. B. Jin, M. G. K. Ong, and D. N. C. Ling, “An automated palmprint recognition system,” Image Vis. Comput., vol. 23, no. 5,
pp. 501–515, 2005.
[10] J. Daugman, “The importance of being random: Statistical principles of
iris recognition,” Pattern Recognit., vol. 36, no. 2, pp. 279–291, 2003.
[11] M. De Marsico, M. Nappi, D. Riccio, and H. Wechsler, “Robust face
recognition for uncontrolled pose and illumination changes,” IEEE
Trans. Syst., Man, Cybern., Syst., vol. 43, no. 1, pp. 149–163, Jan. 2013.
[12] M. R. Diaz, C. M. Travieso, J. B. Alonso, and M. A. Ferrer, “Biometric
system based in the feature of hand palm,” in Proc. 38th Annu. Int.
Carnahan Conf. Security Technol., Albuquerque, NM, USA, 2004,
pp. 136–139.
[13] N. Duta, A. K. Jain, and K. V. Mardia, “Matching of palmprints,” Pattern
Recognit., vol. 23, no. 4, pp. 477–485, 2002.
[14] M. Ekinci and M. Aykut, “Palmprint recognition by applying waveletbased kernel PCA,” J. Comput. Sci. Technol., vol. 23, no. 5, pp. 851–861,
2008.
[15] M. Fang, G. Yue, and Q. Yu, “The study on an application of Otsu
method in canny operator,” in Proc. Int. Symp. Inf. Process. (ISIP), 2009,
pp. 109–112.
[16] L. Fei, Y. Xu, and D. Zhang, “Half-orientation extraction of palmprint
features,” Pattern Recognit. Lett., vol. 69, pp. 35–41, Jan. 2016.
[17] L. Fei, Y. Xu, W. Tang, and D. Zhang, “Double-orientation code
and nonlinear matching scheme for palmprint recognition,” Pattern
Recognit., vol. 49, pp. 89–101, Jan. 2016.
[18] Y. Gao and M. K. H. Leung, “Face recognition using line edge map,”
IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 6, pp. 764–779,
Jun. 2002.
[19] C.-C. Han, H.-L. Cheng, C.-L. Lin, and K.-C. Fan, “Personal authentication using palm-print features,” Pattern Recognit., vol. 36, no. 2,
pp. 371–381, 2003.
[20] L. Hong, Y. Wan, and A. Jain, “Fingerprint image enhancement:
Algorithm and performance evaluation,” IEEE Trans. Pattern Anal.
Mach. Intell., vol. 20, no. 8, pp. 777–789, Aug. 1998.
[21] D.-S. Huang, W. Jia, and D. Zhang, “Palmprint verification based on
principal lines,” Pattern Recognit., vol. 41, no. 4, pp. 1316–1328, 2008.
[22] A. K. Jain and J. Feng, “Latent palmprint matching,” IEEE Trans.
Pattern Anal. Mach. Intell., vol. 31, no. 6, pp. 1032–1047, Jun. 2009.
[23] A. K. Jain, S. Prabhakar, L. Hong, and S. Pankanti, “Filterbankbased fingerprint matching,” IEEE Trans. Image Process., vol. 9, no. 5,
pp. 846–859, May 2000.
[24] W. Jia, R.-X. Hu, Y.-K. Lei, Y. Zhao, and J. Gui, “Histogram of oriented
lines for palmprint recognition,” IEEE Trans. Syst., Man, Cybern., Syst.,
vol. 44, no. 3, pp. 385–395, Mar. 2014.
[25] W. Jia, D.-S. Huang, and D. Zhang, “Palmprint verification based
on robust line orientation code,” Pattern Recognit., vol. 41, no. 5,
pp. 1504–1513, 2008.
[26] W. Kang and Q. Wu, “Pose-invariant hand shape recognition based on
finger geometry,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 44, no. 11,
pp. 1510–1521, Nov. 2014.
[27] A. W. K. Kong, D. Zhang, and M. Kamel, “A survey of palmprint
recognition,” Pattern Recognit., vol. 42, no. 7, pp. 1408–1418, 2009.
[28] A. W. K. Kong, D. Zhang, and W. Li, “Palmprint feature extraction using
2-D Gabor filters,” Pattern Recognit., vol. 36, no. 10, pp. 2339–2347,
2003.
[29] A. W.-K. Kong and D. Zhang, “Competitive coding scheme for palmprint verification,” in Proc. 17th Int. Conf. Pattern Recognit. (ICPR),
vol. 1. Cambridge, U.K., 2004, pp. 520–523.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
PALMA et al.: BIOMETRIC PALMPRINT VERIFICATION: DYNAMICAL SYSTEM APPROACH
[30] K. Krishneswari and S. Arumugam, “Intramodal feature fusion based
on PSO for palmprint authentication,” J. Image Video Process., vol. 2,
no. 4, pp. 435–440, 2012.
[31] A. Kumar, D. C. M. Wong, H. C. Shen, and A. K. Jain, Personal
Verification Using Palmprint and Hand Geometry Biometric (Lecture
Notes in Computer Science). vol. 2688. Berlin, Germany: Springer,
2003, pp. 668–678.
[32] A. Kumar and D. Zhang, “Personal authentication using multiple palmprint representation,” Pattern Recognit., vol. 38, no. 10, pp. 1695–1704,
2005.
[33] S. Y. Kung, S.-H. Lin, and M. Fang, “A neural network approach to
face/palm recognition,” in Proc. IEEE Workshop Neural Netw. Signal
Process., Cambridge, MA, USA, 1995, pp. 323–332.
[34] V. Laxmi, “Palmprint matching using LBP,” in Proc. IEEE Int. Conf.
Comput. Sci. (ICCS), 2012, pp. 110–115.
[35] M. A. L. Vijilious and V. S. Bharathi, “Texture feature extraction
approach to palmprint using nonsubsampled contourlet transform and
orthogonal moments,” Int. J. Future Comput. Commun., vol. 1, no. 3,
pp. 298–301, 2012.
[36] M. K. H. Leung, A. C. M. Fong, and S. C. Hui, “Palmprint verification
for controlling access to shared computing resources,” IEEE Pervasive
Comput., vol. 6, no. 4, pp. 40–47, Oct./Dec. 2007.
[37] F. Li, M. K. H. Leung, and X. Yu, “Palmprint identification using
Hausdorff distance,” in Proc. Int. Workshop Biomed. Circuits Syst.,
Singapore, 2004, pp. 1–6.
[38] G. Lu, D. Zhang, and K. Wang, “Palmprint recognition using eigenpalms
features,” Pattern Recognit. Lett., vol. 24, nos. 9–10, pp. 1463–1467,
2003.
[39] Y.-T. Luo et al., “Local line directional pattern for palmprint recognition,” Pattern Recognit., vol. 50, pp. 26–44, Feb. 2016.
[40] A. Nigam and P. Gupta, “Palmprint recognition using geometrical and
statistical constraints,” in Proc. 2nd Int. Conf. Soft Comput. Problem
Solving (SocProS), Jaipur, India, 2014, pp. 1303–1315.
[41] J. S. Noh and K. H. Rhee, “Palmprint identification algorithm using Hu
invariant moments and Otsu binarization,” in Proc. 4th Annu. ACIS Int.
Conf. Comput. Inf. Sci., 2005, pp. 94–99.
[42] G. K. O. Michael, T. Connie, and A. B. J. Teoh, “Touch-less palm
print biometrics: Novel design and implementation,” Image Vis. Comput.,
vol. 26, no. 12, pp. 1551–1560, 2008.
[43] N. Otsu, “A threshold selection method from gray-level histograms,”
IEEE Trans. Syst., Man, Cybern., vol. SMC-9, no. 1, pp. 62–66,
Jan. 1979.
[44] D. Palma, P. L. Montessoro, G. Giordano, and F. Blanchini, “A dynamic
algorithm for palmprint recognition,” in Proc. IEEE Conf. Commun.
Netw. Security (IEEE CNS) 1st Workshop Security Privacy Cybermatics
(SPiCy), Florence, Italy, 2015, pp. 623–626.
[45] C. Phromsuthirak, S. Suwan, A. Sanpanich, and C. Pintavirooj, “Hand
shape identification using palmprint alignment based on intrinsic local
affine-invariant fiducial points,” in Proc. Biomed. Eng. Int. Conf.
(BMEiCON), Fukuoka, Japan, 2014, pp. 1–5.
[46] X. Pin and Q.-Q. Ruan, “Palmprint recognition using Gabor
feature-based (2D)2 PCA,” Neurocomputing, vol. 71, nos. 13–15,
pp. 3032–3036, 2008.
[47] R. Raghavendra, B. Dorizzi, A. Rao, and G. H. Kumar, “PSO versus
adaboost for feature selection in multimodal biometrics,” in Proc. 3rd
IEEE Int. Conf. Biometrics Theory Appl. Syst., Washington, DC, USA,
2009, pp. 238–244.
[48] S. Ribaric, I. Fratric, and K. Kis, “A biometric verification system based
on the fusion of palmprint and face features,” in Proc. 4th Int. Symp.
Image Signal Process. Anal., Zagreb, Croatia, 2005, pp. 12–17.
[49] R. Sanchez-Reillo, C. Sanchez-Avilla, and A. Gonzalez-Marcos,
“Biometric identification through hand geometry measurements,” IEEE
Trans. Pattern Anal. Mach. Intell., vol. 22, no. 10, pp. 1168–1171,
Oct. 2000.
[50] L. Shengcai, A. K. Jain, and S. Z. Li, “Partial face recognition:
Alignment-free approach,” IEEE Trans. Pattern Anal. Mach. Intell.,
vol. 35, no. 5, pp. 1193–1205, May 2013.
[51] Z. Sun, Y. Wang, T. Tan, and J. Cui, “Robust direction estimation of
gradient vector field for iris recognition,” in Proc. 17th Int. Conf. Pattern
Recognit., Cambridge, U.K., 2004, pp. 783–786.
[52] Z. Sun, T. Tan, Y. Wang, and S.Z. Li, “Ordinal palmprint represention for
personal identification [represention read representation],” in Proc. IEEE
Comput. Soc. Conf. Comput. Vis. Pattern Recognit., vol. 1. San Diego,
CA, USA, 2005, pp. 279–284.
11
[53] J. A. Wincy and J. G. C. Chandran “Palmprint recognition using PCF
and SURF,” Int. J. Adv. Res. Comput. Sci. Softw. Eng., vol. 3, no. 10,
pp. 996–1001, 2013.
[54] X. Wu, K. Wang, and D. Zhang, “Fuzzy directional element energy feature (FDEEF) based palmprint identification,” in Proc. Int. Conf. Pattern
Recognit., vol. 1. Quebec City, QC, Canada, 2002, pp. 95–98.
[55] X. Wu, K. Wang, and D. Zhang, “Line feature extraction and matching
in palmprint,” in Proc. 2nd Int. Conf. Image Graph., 2002, pp. 583–590.
[56] X. Wu, D. Zhang, K. Wang, and B. Huang, “Palmprint classification
using principal lines,” Pattern Recognit., vol. 37, no. 10, pp. 1987–1998,
2004.
[57] X. Wu, K. Wang, and D. Zhang, HMMs Based Palmprint Identification
(Lecture Notes in Computer Science), vol. 3072. Berlin, Germany:
Springer, 2004, pp. 775–781.
[58] X. Wu, D. Zhang, and K. Wang, “Palm line extraction and matching
for personal authentication,” IEEE Trans. Syst., Man, Cybern. A, Syst.,
Humans, vol. 36, no. 5, pp. 978–987, Sep. 2006.
[59] J. You, W. Li, and D. Zhang, “Hierarchical palmprint identification
via multiple feature extraction,” Pattern Recognit., vol. 35, no. 4,
pp. 847–859, 2002.
[60] D. Zhang, A. W.-K. Kong, J. You, and M. Wong, “Online palmprint
identification,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 25, no. 9,
pp. 1041–1050, Sep. 2003.
[61] B. Zhang, W. Li, P. Qing, and D. Zhang, “Palm-print classification by
global features,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 43, no. 2,
pp. 370–378, Mar. 2003.
[62] D. Zhang and W. Shu, “Two novel characteristics in palmprint verification: Datum point invariance and line feature matching,” Pattern
Recognit., vol. 32, no. 4, pp. 691–702, 1999.
[63] Z.J. Zia-uddin, Z. Jan, J. Ahmad, and A. Abbasi, “A novel technique
for principal lines extraction in palmprint using morphological top-hat
filtering,” World Appl. Sci. J., vol. 31, no. 12, pp. 2010–2014, 2014.
[64] PolyU Palmprint Database, Biometric Res. Center Hong Kong
Polytech. Univ., Hong Kong, 2003. [Online]. Available:
http://www.comp.polyu.edu.hk/∼biometrics/
[65] CASIA Palmprint Image Database, Chin. Acad. Sci. Inst. Autom.,
Beijing, China, 2005. [Online]. Available: http://biometrics.idealtest.org/
David Palma received the B.Sc. degree in
information engineering and management and the
M.Sc. degree in electronic engineering from the
University of Udine, Udine, Italy, in 2014 and
2017, respectively, where he is currently pursuing
the Ph.D. degree with the Polytechnic Department
of Engineering and Architecture.
From 2014 to 2017, he was with the Department
of Electrical, Mechanical and Management
Engineering, University of Udine, where he was
a Research Assistant in biometric authentication,
cybersecurity, and pervasive computing. His current research interests include
biometric authentication, pervasive computing, and cybersecurity.
Pier Luca Montessoro was born in Turin, Italy, in
1961. He received the Dr.Eng. degree in electronic
engineering from the Polytechnic of Turin, Turin, in
1986.
He is currently a Full Professor of computer science with the University of Udine, Udine, Italy.
He has been with the Italian National Council for
Scientific Research, Rome, Italy, and a Scientific
Consultant with Digital Equipment Corporation,
Maynard, MA, USA (currently Compaq) in the field
of simulation for very large scale integration design.
His research interests, after several years spent on CAD systems for digital
circuits design and on multimedia systems for e-learning. His current research
interests include computer networks, ICT security and pervasive computing,
in particular distributed controls and algorithms for agents-based systems.
He has been the Chair and an Organizer of the WCC 2013 workshop “International Workshop on Cloud Convergence: Challenges for Future
Infrastructures and Services,” hosted in the IEEE ICC conference and chair
of the 30th edition of the Didamatica conference, held in Udine.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
12
Giulia Giordano received the B.Sc. and M.Sc.
degrees (summa cum laude) and the Ph.D. degree
from the University of Udine, Udine, Italy, in 2010,
2012, and 2016, respectively.
She was with the Department of Automatic
Control, Lund University, Lund, Sweden. She is
currently an Assistant Professor with the Delft
Center for Systems and Control, Delft University
of Technology, Delft, The Netherlands. She visited the California Institute of Technology, Pasadena,
CA, USA, in 2012 and the University of Stuttgart,
Stuttgart, Germany, in 2015. Her current research interests include analysis
of biological systems and the control of networked systems.
Dr. Giordano was a recipient of the EECI Ph.D. Award 2016 for her doctoral
thesis and the NAHS Best Paper Award 2017.
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
Franco Blanchini was born in Legnano, Italy, in
1959. He received the Laurea degree in electrical
engineering from the University of Trieste, Trieste,
Italy, in 1984.
He is the Director of the Laboratory of System
Dynamics, University of Udine, Udine, Italy. He has
co-authored the book entitled Set Theoretic Methods
in Control (Birkhäuser).
Mr. Blanchini was a recipient of the 2001 ASME
Oil and Gas Application Committee Best Paper
Award, the 2017 NAHS Best Paper Award, and the
2002 IFAC Prize Survey Paper Award and the High Impact Paper Award
2017 for his article Set Invariance in Control—A Survey, Automatica, 1999.
He was an Associate Editor of Automatica, from 1996 to 2006 and the IEEE
T RANSACTIONS ON AUTOMATIC C ONTROL, from 2012 to 2016. Since 2017,
he has been an Associate Editor of Automatica. He is a Senior Editor of the
IEEE C ONTROL S YSTEMS L ETTERS. He has been involved in the organization of several international events: he was the Program Vice-Chairman of
IEEE CDC-ECC 2005, Seville, Spain, the Program Vice-Chairman of IEEE
CDC 2008, Cancun, Mexico, the Program Chairman of IFAC ROCOND 2012,
Aalborg, Denmark, and the Program Vice-Chairman of IEEE CDC 2013,
Florence, Italy.