Conference PaperPDF Available

Palmprint Recognition Under Unconstrained Scenes

Authors:

Abstract and Figures

This paper presents a novel real-time palmprint recognition system for cooperative user applications. This system is the first one achieving non-contact capturing and recognizing palmprint images under unconstrained scenes. Its novelties can be described in two aspects. The first is a novel design of image capturing device. The hardware can reduce influences of background objects and segment out hand regions efficiently. The second is a process of automatic hand detection and fast palmprint alignment, which aims to obtain normalized palmprint images for subsequent feature extraction. The palmprint recognition algorithm used in the system is based on accurate ordinal palmprint representation. By integrating power of the novel imaging device, the palmprint preprocessing approach and the palmprint recognition engine, the proposed system provides a friendly user interface and achieves a good performance under unconstrained scenes simultaneously.
Content may be subject to copyright.
Y. Yagi et al. (Eds.): ACCV 2007, Part II, LNCS 4844, pp. 1–11, 2007.
© Springer-Verlag Berlin Heidelberg 2007
Palmprint Recognition Under Unconstrained Scenes
Yufei Han, Zhenan Sun, Fei Wang, and Tieniu Tan
Center for Biometrics and Security Research
National Laboratory of Pattern Recognition, Institute of Automation
Chinese Acdamey of Sciences
P.O.Box 2728, Beijing, P.R. China, 100080
{yfhan,znsun,fwang,tnt}@nlpr.ia.ac.cn
Abstract. This paper presents a novel real-time palmprint recognition system
for cooperative user applications. This system is the first one achieving non-
contact capturing and recognizing palmprint images under unconstrained
scenes. Its novelties can be described in two aspects. The first is a novel design
of image capturing device. The hardware can reduce influences of background
objects and segment out hand regions efficiently. The second is a process of
automatic hand detection and fast palmprint alignment, which aims to obtain
normalized palmprint images for subsequent feature extraction. The palmprint
recognition algorithm used in the system is based on accurate ordinal palmprint
representation. By integrating power of the novel imaging device, the palmprint
preprocessing approach and the palmprint recognition engine, the proposed
system provides a friendly user interface and achieves a good performance
under unconstrained scenes simultaneously.
1 Introduction
Biometrics technology identifies different people by their physiological and
behavioral differences. Compared with traditional security authentication approaches,
such as key or password, biometrics is more accurate, dependable and difficult to be
stolen or faked. In the family of biometrics, palmprint is a novel but promising
member. Large region of palm supplies plenty of line patterns which can be easily
captured in a low resolution palmprint image. Based on those line patterns, palmprint
recognition can achieve a high accuracy of identity authentication.
In previous work, there are several successful recognition systems proposed for
practical use of palmprint based identity check [1][2][3], and the best-known is
developed by Zhang et al [1]. During image capturing, users are required to place
hands on the plate with pegs controlling displacement of hands. High quality
palmprint images are then captured by a CCD camera fixed in a semi-closed
environment with uniform light condition. To alignment captured palmprint images, a
preprocessing algorithm [2] is adopted to correct rotation of those images and crop
square ROI (regions of interests) with the same size. Detail about this system can be
found in [2]. Besides, Connie et al proposed a peg-free palmprint recognition system
[3], which captures palmprint images by an optical scanner. Subjects are allowed to
place their hand more freely on the platform of the scanner without pegs. As a result,
2 Y. Han et al.
palmprint images with different sizes, translations and rotation angles are obtained.
Similar as in [2], an alignment process is involved to obtain normalized ROI images.
However, efficient as they are, there are still some limitations. Firstly, some users
may feel uncomfortable with pegs to restrict hands during capturing images.
Secondly, even without pegs, subjects’ hands are required to contact plates of devices
or platforms of scanners, which is not hygienic enough. Thirdly, semi-closed image
capturing devices usually increase volume of recognition systems, which makes them
not convenient for portable use. Thus, it’s necessary to improve design of the
HCI(human-computer interface), in order to make the whole system easy-to-use.
Recently, active near infrared imagery (NIR) technology has received more and
more attention in face detection and recognition, as seen in [4]. Given a near
infrared light source shining objects in front of cameras, intensity of reflected NIR
light is attenuated at a large scale with distance between objects and the light
source increasing. This property provides us a promising solution to eliminate
affection of backgrounds when palmprint images are captured under unconstrained
scenes. Based on the technology, in this paper, we propose a novel real-time
palmprint recognition system. It’s designed to localize and obtain normalized
palmprint images under clutter scenes conveniently. The main contributions are as
followings: First, we present a novel design of portable image capturing device,
which mainly consists of two parallel placed web cameras. One is used for active
near infrared imagery to localize hand regions. The other one captures
corresponding palmprint images in visible light, preparing for further feature
extraction. Second, we present a novel palmprint preprocessing algorithm, utilizing
color and shape information of hands for fast and effective hand region detection,
rotation correction and localization of central palm region. So far as we know,
there is no similar work reported in previous literatures.
The rest of paper is organized as follows: Section 2 presents a description of the
whole architecture of the recognition system. In Section 3, the design of human
computer interface of the system is described in detail. Section 4 introduces ordinal
palmprint representation briefly. Section 5 evaluates the performance of the system.
Finally, in Section 6, we conclude the whole paper.
2 System Overview
We adopt a common PC with Intel Pentium4 3.0Ghz and 1G RAM as the
computation platform. Based on it, the recognition system is implemented using
Microsoft Visual C++ 6.0. It consists of five main modules, as shown in Fig.1. After
starting the system, users are required to open their hands in a natural manner and
place palm regions toward the imaging device at a certain distance between 35 cm
and 50 cm from cameras. Surfaces of palms are approximately orthogonal to the
optical axis of cameras. In-plane rotation of hands is restricted between -15 degree to
15 degree deviated from vertical orientation. The imaging device then captures two
images for each hand by two cameras placed in parallel respectively. One is a NIR
hand image with active NIR lighting, the other is a color hand image with background
objects, obtained with normal environment lighting condition. Both of them contain
complete hand region, see in Fig.2. After that, an efficient palmprint preprocessing
Palmprint Recognition Under Unconstrained Scenes 3
algorithm is performed on the two captured images to obtain one normalized
palmprint image quickly, which makes use of both shape and skin color information
of hands. Finally, robust palmprint feature templates are extracted from the
normalized image using the ordinal code based approach [5]. Fast hamming distance
calculation is applied to measure dissimilarity between two feature templates. An
example of the whole recognition process could be seen in the supplementary video of
this paper.
3 Smart Human-Computer Interface
HCI of the system mainly consists of two parts, image capturing hardware and
palmprint preprocessing procedure, as shown in Fig.1. Considering a hand image
captured under an unconstrained scene, unlike those captured by devices in
[1][2][3], there are not only a hand region containing palmprint patterns, but also
background objects of different shapes, colors and positions, as denoted in Fig.2.
Even within the hand, there still exits rotation, scale variation and translation of
palmprint patterns due to different hand displacements. Thus, before further
palmprint feature encoding, HCI should localize the candidate hand region and
extract a normalized ROI (region of interest), which contains palmprint features
without much geometric deformations.
3.1 Image Capturing Device
Before palmprint alignment, it is necessary to segment hand regions from
unconstrained scenes. This problem could be solved by background modeling and
subtraction or labeling skin color region. However, both methods suffer from
unconstrained backgrounds or varying light conditions. Our design of imaging device
aims to solve the problem in a sensor level, in order to localize foreground hand
regions more robustly by simple image binarization.
The appearance of the image capturing device is shown in Fig.2(a). This device has
two common CMOS web cameras placed in parallel. We mount near infrared (NIR)
light-emitting diodes on the device evenly distributed around one camera, similar as
in [4], so as to provide straight and uniform NIR lighting. Near infrared light emit by
those LEDs have a wavelength of 850 nm. In a further step, we make use of a band
pass optical filter fixed on the camera lens to cut off lights with all the other
wavelengths except 850nm. Most of environment lights are cut off because their
wavelengths are less than 700nm. Thus, lights received by the camera only consist of
reflected NIR LED lights and NIR components in environment lights, such as lamp
light and sunlight, which are much weaken than the NIR LED lights. Notably,
intensities of reflected NIR LED lights are in the inverse proportion to high-order
terms of the distance between object and the camera. Therefore, assuming a hand is
the nearest one among all objects in front of the camera during image capturing,
intensities of the hand region in the corresponding NIR image should be much larger
than backgrounds. As a result, we can further segment out the hand region and
eliminate background by fast image binarization, as denoted in Fig.2(b). The other
4 Y. Han et al.
camera in the device captures color scene images, obtaining clear palmprint patterns
and reserving color information of hands. An optical filter is fixed on the lens of this
camera to filter out infrared components in the reflected lights, which is applied
widely in digital camera to avoid red-eye. The two cameras work simultaneously. In
our device, resolution of both cameras is 640*480. Fig.2(b) lists a pair of example
images, captured by the two cameras at the same time. The upper one is the color
image. The bottom one is the NIR image. The segmentation result is shown in the
upper row of Fig.2(c). In order to focus on hand regions with a proper scale in further
processing, we adopt a scale selection on binary segmentation results to choose
candidate foreground regions. The criterion of selection grounds on a fact that area of
a hand region in a NIR image is larger if the hand is nearer to the camera. We label all
connected binary foreground after segmentation and calculate area of each connected
component, then choose those labeled regions with their areas varying in a
predefined narrow range as the candidate foreground regions, like the white region
shown in the image at the bottom of Fig.2(c).
Fig. 1. Flowcharts of the system
Fig. 2. (a) Image capturing device (b) Pair-wise color and NIR image (c) Segmented fore
ground and candidate foreground region selection
Palmprint Recognition Under Unconstrained Scenes 5
3.2 Automated Hand Detection
Hand detection is posed as two-class problem of classifying the input shape pattern
into hand-like and non-hand class. In our system, a cascade classifier is trained to
detect hand regions in binary foregrounds, based on works reported in [6]. In [6],
Eng-Jon Ong et al makes use of such classifier to classify different hand gestures. In
our application, the cascade classifier should be competent for two tasks. Firstly, it
should differentiate shape of open hand from all the other kinds of shapes. Secondly,
it should reject open hands with in-plane rotation angle deviating out of the restricted
range. To achieve this goals, we construct a positive dataset containing binary open
left hands at first, such as illustrated in Fig.3(a). In order to make the classifier
tolerate certain in-plane rotation, the dataset consists of left hands with seven discrete
rotation angles, sampled every 5 degree from -15 degree to 15 degree deviated from
vertical orientation, a part of those binary hands are collected from [11]. For each
angle, there are about 800 hand images with slight postures of fingers, also shown in
Fig.3(a). Before training, all positive data are normalized into 50*35 images. The
negative dataset contains two parts. One consists of binary images containing non-
hand objects, such as human head, turtles and cars, partly from [10]. The other
contains left hands with rotation angle out of the restricted range and right hands with
a variety of displacements. There are totally more than 60,000 negative images.
Fig.3(b) shows example negative images. Based on those training data, we use Float
AdaBoost algorithm to select most efficient Haar features to construct the cascade
classifier, same as in [6]. Fig.3(c) shows the most six efficient Haar features obtained
after training. We see that they represent discriminative shape features of left open
hand. During detection, rather than exhaustive search across all positions and scales in
[6], we perform the classifier directly around the candidate binary foreground regions
Fig. 3. (a) Positive training data (b) Negative training data (c) Learned efficient Haar features
(d) Detected hand region
6 Y. Han et al.
to search for open left hands with a certain scale. Therefore, we can detect different
hands with a relative stable scale, which reduces influence of scale variations on
palmprint patterns. Considering mirror symmetry between left and right hands, to
detect right hands, we just perform symmetry transform on the images and apply the
classifier by the same way on the flipped images. Fig.3(d) shows results of detection.
Obtaining detected hand, all the other non-hand connected regions are removed from
binary hand images. The whole detection can be finished within 20 ms.
3.3 Palmprint Image Alignment
Palmprint alignment procedure eliminates rotation and translation of palmprint
patterns, in order to obtain normalized ROI. Most alignment algorithms calculate
rotation angles of hands by localizing key contour points in gaps between fingers
[2][3]. However, in our application, different finger displacements may change local
contours and make it difficult detect gap regions, as denoted in Fig.4. To solve this
problem, we adopt a fast rotation angle estimation based on moments of hand shape.
Given R is the detected hand region in a binary foreground image. Its orientation
θ
can be estimated by calculating its moments [7]:
1,1
2,0 0, 2
2
1arctan( )
2
μ
θμμ
= (1)
p,q
μ
(p,q=0,1….) is (p,q) order central moments, which is represented as :
,
11
()(),(,)R
pq
yxy xy
pq
x
xxyyxy
NN
μ
=− −
∑∑ ∑∑ ∑∑ (2)
Compared with key point detection, moments are calculated based on the whole
hand region rather than only contour points. Thus, it is more robust to local changes in
contours. To reduce computation cost, the original binary image is down-sampled to a
160*120 one. Those moments are then calculated on the down-sampled version. After
obtaining rotation angle
θ
, the hand region is rotated by -
θ
degree to get vertical
oriented hands, see in Fig.4. Simultaneously, the corresponding color image is also
rotated by -
θ
, in order to make sure consistency of hand orientations in both two
images.
In a further step, we locate central palm region in a vertical oriented open hand by
analyzing difference of connectivity between the palm region and the finger region.
Although shape and size of hands vary a lot, a palm region of each hand should be
like a rectangle. Compared with it, stretched fingers don’t form a connective region as
palm. Based on this property, we employ an erosion operation on the binary hand
image to remove finger regions. The basic idea behind this operation is run length
code of binary image. We perform a raster scanning on each row to calculate the
maximum length W of connective sequences in the row. Any row with its W less than
threshold K1 should be eroded. After all rows are scanned, a same operation is
performed on each column. As a result, columns with their maximum length W less
than K2 are removed. Finally, a rectangular palm region is cropped from the hand.
Coordinates (xp,yp )of its central point is derived as localization result. In order to
Palmprint Recognition Under Unconstrained Scenes 7
cope with varying sizes of different hands, we choose values of K1 and K2 adaptively.
Before row erosion, distance between each point in the hand region and nearest edge
point is calculated by a fast distance transform. The central point of hand is defined as
the one with the largest distance value. Assuming A is the maximum length of
connective sequences in the row passing through the central point, K1 is defined as
follows:
K1 = A * p% (3)
p is a pre-defined threshold. K2 is defined in the same way:
K2 = B * q% (4)
B is the maximum length of connective sequences in the column passing through the
central point after row erosion. q is another pre-defined threshold. Compared with
fixed value, adaptive K1 and K2 lead to more accurate location of central palm regions,
as denoted in Fig.5(b). Fig.5(a) denotes the whole procedure of erosion.
Due to visual disparity between two cameras in the imaging device, we can not use
(xp,yp ) to localize ROI in corresponding color images directly. Although visual
disparity can be estimated by a process of 3D scene reconstruction, this approach may
lead to much computation burden on the system. Instead, we apply a fast
correspondence estimation based on template matching. Assuming C is a color hand
image after rotation correction, we convert C into a binary image M by setting all
pixels with skin color to 1, based on the probability distribution model of skin color in
RGB space [8]. Given the binary version of the corresponding NIR image, with a
hand region S locating at (xn,yn), a template matching is conducted as in Eq.5, also as
denoted in Fig.6:
(,) [ ( , ) (,)],(,)
xy
fmn Mx my n Sxy xy S=++⊕ ∈
∑∑ (5)
is bitwise AND operator. f( , ) is a matching energy function. (m,n) is a candidate
position of the template. The optimal displacement (xo,yo) of hand shape S in M is
defined as the candidate position where the matching energy achieves its maximum.
The central point (xc,yc) of palm region in C can be estimated by following equations:
cpon
cpon
xxxx
yyyy
=+
=+
(6)
Fig. 4. Rotation correction
8 Y. Han et al.
Fig. 5. (a) Erosion procedure (b) Erosion with fixed and adaptive thresholds
With (xc,yc) as its center, one 128*128 sub-image is cropped from C as ROI, which is
then converted to gray scale image for feature extraction.
Fig. 6. Translation estimation
4 Ordinal Palmprint Representation
In previous work, the orthogonal line ordinal feature (OLOF) [5] provides a compact
and accurate representation of negative line features in palmprints. The orthogonal
line ordinal filter [5] F(x,y,θ) is designed as follows:
(,,) (,,) (,, /2)Fxy Gxy Gxy
θθθπ
=−+
(7)
Palmprint Recognition Under Unconstrained Scenes 9
22
x
cos sin sin cos
(,,) exp[( ) ( )]
y
xy xy
Gxy
θθ θθ
θδδ
+−+
=− − (8)
G(x,y,θ) is a 2D anisotropic Gaussian filter, and θ is the orientation of the Gaussian
filter. The ratio between δx and δy is set to be larger than 3, in order to obtain a
weighted average of a line-like region. In each local region in a palmprint image,
three such ordinal filters, with orientations of 0, π/6, π/3 are used in convolution
process on the region. The filtering result is then encoded into 1 or 0 according to
whether its sign is positive or negative. Thousands of ordinal codes are concatenated
into a feature template. Similarity between two feature templates is measured by a
normalized hamming distance, which ranges between 0 and 1. Further details can be
found in [5].
5 System Evaluation
Performance of the system is evaluated in terms of verification rate [9], which is
obtained through one-to-one image matching. We collect 1 00 normalized palmprint
ROI images from 0 subjects using the system, with 10 images for each hand. Fig.7
illustrates six examples of ROI images. During the test, there are totally 5,400 intra-
class comparisons and 714,000 inter-class comparisons. Although recognition
accuracy of the system lies on effectiveness of both alignment procedure of HCI and
the palmprint recognition engine, the latter is not the focus of this paper. Thus we
don’t involve performance comparisons between the ordinal code and other state-of-
the-art approaches. Fig.8 denotes distributions of genuine and imposter. Fig.9 shows
corresponding ROC curve. The equal error rate [9] of the verification test is 0.54%.
From experimental results, we can see that ROI regions obtained by the system are
suitable for palmprint feature extraction and recognition. Besides, we also record time
cost for obtaining one normalized palmprint image using the system. It includes time
for image capturing, hand detection and palmprint alignment. The average time cost is
1.2 seconds. Thus, our system can be competent for point-of-sale identity check.
Fig. 7. Six examples of ROI images
10 Y. Han et al.
Fig. 8. Distributions of genuine and imposter
Fig. 9. ROC curve of the verification test
6 Conclusion
In this paper, we have proposed a novel palmprint recognition system for cooperative
user applications, which achieves a real-time non-contact palmprint image capturing
and recognition directly under unconstrained scenes. Through design of the system,
we aim to provide more convenient human-computer interface and reduce restriction
on users during palmprint based identity check. The core of HCI in the system
consists of a binocular image device and a novel palmprint preprocessing algorithm.
The former delivers a fast hand region segmentation based on NIR imaging
technology. The latter extracts normalized ROI from hand regions efficiently based
on shape and color information of human hands. Benefiting further from the powerful
recognition engine, the proposed system achieves accurate recognition and convenient
use at the same time. As far as we know, this is the first attempt to solve the problem
of obtaining normalized palmprint images directly from clutter backgrounds.
However, accurate palmprint alignment has not been well addressed in the
proposed system. In our future work, it’s an important issue to improve the
performance of the system by reducing alignment error in a further step. In addition,
Palmprint Recognition Under Unconstrained Scenes 11
we should improve the imaging device to deal with influence of NIR component in
environment light, which varies much in practical use.
Acknowledgments. This work is funded by research grants from the National Basic
Research Program (Grant No.2004CB318110), the Natural Science Foundation of
China (Grant No.60335010, 60121302, 60275003, 60332010, 69825105,60605008)
and the Chinese Academy of Sciences.
References
1. Zhang, D., Kong, W.K., You, J., Wong, M.: Online Palmprint Identification. IEEE Trans
on PAMI 25(9), 1041–1050 (2003)
2. Kong, W.K.: Using Texture Analysis in Biometric Technology for Personal Identification,
MPhil Thesis, http://pami.uwaterloo.ca/ cswkkong/Sub_Page/Publications.htm
3. Connie, T., Jin, A.T.B., Ong, M.G.K., Ling, D.N.C.: Automated palmprint recognition
system. Image and Vision Computing 23, 501–515 (2005)
4. li, S.Z., Chu, R.F., Liao, S.C., Zhang, L.: Illumination invariant Face Recognition using
Near- Infrared Images. IEEE Trans on PAMI 29(4), 627–639 (2007)
5. Sun, Z.N., Tan, T.N., Wang, Y.H., Li, S.Z.: Ordinal Palmprint Representation for Personal
Identification. Proc. of IEEE CVPR 2005 1, 279–284 (2005)
6. Ong, E., Bowden, R.: A Boosted Classifier Tree for Hand Shape Detection. In: Proc. of
International Conference on Automatic Face and Gesture Recognition, pp. 889–894 (2004)
7. Jain, A.K.: Fundamentals of Digital Image Processing, vol. 07458, p. 392. Prentice Hall,
Upper Saddle River, NJ
8. Jones, M.J., Rehg, J.M.: Statistical Color Models with Application to Skin Color
Detection. International Journal of Computer Vision 46(1), 81–96 (2002)
9. Daugman, J., Williams, G.: A Proposed Standard for Biometric Decidability. In: Proc. of
CardTech/SecureTech Conference, Atlanta, GA, pp. 223–234 (1996)
10. http://www.cis.temple.edu/ latecki/TestData mpeg7shapeB.tar.gz
11. UST Hand Image database, http://visgraph.cs.ust.hk/Biometrics/Visgraph_web/
index.html
... Compared to monocular systems, dual-camera systems can capture more details and achieve high-quality images. There are two types of dual-camera systems: one utilizes bimodal cameras, and the other uses two identical RGB cameras: 1) the bimodal system captures palmprint and palm vein images separately and performs living palmprint recognition by fusing the two modalities [17], [18] and 2) the RGB system captures two RGB palm images and reconstructs the 3-D information of the palm, e.g., previous studies [19] and [20] applied the stereo vision technique to touchless palmprint verification to correct palm plane tilt. However, the keypoint detection and matching processes used in the second type of systems may be sensitive to image blur and ambient light conditions. ...
... After bimodal images are captured, image localization is performed and aligned bimodal ROI images are extracted. Generally, current palm alignment methods are based on keypoint matching [21], template matching [17], and shape remapping [22]. For bimodal images, the challenge is that the same point has different textures in the two images because they are generated by different optical spectrums. ...
... For bimodal images, the challenge is that the same point has different textures in the two images because they are generated by different optical spectrums. Han et al. [17] used threshold-based and skin-color-based segmentation methods in the IR and RGB palm images, respectively, to generate bimodal palm regions. Then, they aligned the two binarized palm region images using a matching energy function. ...
Article
Recently, contactless bimodal palmprint recognition technology has attracted increased attention due to the COVID-19 pandemic. Many dual-camera-based sensors have been proposed to capture palm vein and palmprint images synchronously. However, translations between captured palmprint and palm vein images differ depending on the distance between the hand and the sensors. To address this issue, we designed a low-cost method to align the bimodal palm regions for current dual-camera systems. In this study, we first implemented a contactless palm image acquisition device with a dual-camera module and a single-point time of flight (TOF) ranging sensor. Using this device, we collected a dataset named DCPD under different distances and light source intensities from 271 different palms. Then, a bimodal palm image alignment method is proposed based on the imaging and ranging models. After the system model is calibrated, the translation between the visible light and infrared light palm regions can be estimated quickly based on the palm distance. Finally, we designed a convolutional neural network (CNN) to effectively extract the fine- and coarse-grained palm features. Compared to widely used existing methods, the proposed networks achieved the lowest equal error rate (EER) on the Tongji, IITD, and DCPD datasets, and the average time cost of the system to perform one-time identification is approximately 0.15 s. The experimental results indicate that the proposed methods achieved high efficiency and comparable accuracy. In addition, the system’s EER and rank-1 on the DCPD dataset were 0.304% and 98.66%, respectively.
... Even if the hand and background can be separated by binarisation, the extracted contour may not include the valleys of the fingers when the fingers are closed, and the extraction of the palm region may fail since the keypoints cannot be detected. Several methods for contactless palmprint recognition have been proposed for extracting palmprint regions from hand images [6,[14][15][16], however, these methods also require binarisation of the hand image and detection of finger valleys, and thus they fail to extract palm regions for the same reason. Therefore, the problem to be considered is how to extract a hand region from an input image to utilise the Zhang's method in the contactless situation as well as in the contact situation. ...
... So far, several methods for extracting palm regions have been proposed to improve the accuracy of palmprint recognition [6,13,14,16]. These methods include separating the hand from the background by binarisation and detecting keypoints in the valleys of the fingers. ...
... Therefore, the accuracy of palm region extraction depends on the illumination changes during the acquisition and the complexity of the background. In the following, we describe (a) Zhang et al.'s method [13], which is a standard method, (b) Leng et al.'s method [16], (c) Han et al.'s method [14] and (d) Ito et al.'s method [6], which are methods for contactless palmprint recognition. ...
Article
Full-text available
Abstract Extracting a palm region with fixed location from an input hand image is a crucial task for palmprint recognition to realise reliable person authentication under contactless and unconstrained conditions. A palm region can be extracted from the fixed location using the gaps between fingers. An accurate and robust hand segmentation method is indispensable to extract a palm region from an image with complex background taken under various environments. In this study, HandSegNet, which is a hand segmentation method using Convolutional Neural Network (CNN) for contactless palmprint recognition, is proposed. HandSegNet employs a new CNN architecture consisting of an encoder–decoder model with a pyramid pooling module. Through performance evaluation using a set of synthesised hand images, HandSegNet exhibited the best segmentation results of 98.90% and 93.20% for accuracy and intersection over union, respectively. The effectiveness of HandSegNet in contactless palmprint recognition through experiments using a set of synthesised images of hand images is also demonstrated. Comparing the performance of palmprint recognition using three conventional methods and HandSegNet for palm region extraction, the proposed method has the lowest equal error rate of 4.995%, demonstrating its effectiveness in palm region extraction for contactless palmprint recognition.
... Touchless acquisition and 3D models. A few methods are hand geometry [3], palm-print [4], finger geometry [5], and knuckle-print [6]. To reduce acquisition constraints authors have used Touchless manner. ...
Article
Full-text available
The exponential rise in software computing and hardware technologies has broadened the horizon for different applications in decision making to make human life efficient. Among all the major demands, security systems have always been the dominant one to ensure authenticity of data, source or certain activity. Fingerprint technology has gained wide-spread attention for personalized data, resource or activity accesses authentication. Though, numerous methods have been developed for fingerprint detection and identification, the local input environment, data suitability, distortion and hardware dependency have been the challenge to yield optimal performance. On contrary, the possibilities of touchless 3D-fingerprint identification systems have attracted scientific communities due to ease of implementation, reduced dependency on local environment and sensing hardware. In this paper deep features-based Touchless 3D-Fingerprint Classification System is proposed. In this model a transfer deep-learning model AlexNet-CNN is used for deep feature extraction and classification, which obtains 4096 dimensional deep features. The proposed approach achieves a classification accuracy of 90.20%.
... 1. Capture the palm with a unicolor backboard [4,6,8,12,21,33] 2. Employ an IR camera or a depth camera to capture an IR image or a depth image to assist palm segmentation [34,10,30,35] 3. Enhance the contrast of the foreground and background by setting a strong light source intensity and a short exposure time Their target is enhancing the contrast of the palm region and the background. For example, in [28], the mobile phone's built-in LED flash is utilized for palm segmentation. ...
Chapter
Full-text available
So far, there exist many publicly available palmprint databases. However, not all of them have provided the corresponding region of interest (ROI) images. If everyone uses their own extracted ROI images for performance testing, the final accuracy is not strictly comparable. Since ROI localization is the critical stage of palmprint recognition. The location precision has a significant impact on the final recognition accuracy, especially in unconstrained scenarios. This problem has limited the applications of palmprint recognition. However, many currently published surveys only focus on feature extraction and classification methods. Throughout these years, many new ROI localization methods have been proposed. In this chapter, we will group the existing ROI localization methods into different categories, analyze their basic ideas, reproduce some of the codes, make comparisons of their performances, and provide further directions. We hope this could be a useful reference for further research.
Article
Unconstrained palmprint images have shown great potential for recognition applications due to their lower restrictions regarding hand poses and backgrounds during contactless image acquisition. However, they face two challenges: 1) Unclear palm contours and finger-valley points of unconstrained palmprint images make it difficult to locate landmarks to crop the palmprint region of interest (ROI), and 2) large intra-class diversities of unconstrained palmprint images hinder the learning of intra-class-invariant palmprint features. In this paper, we propose to directly extract the complete palmprint region as the ROI (CROI) using the detection-style CenterNet without requiring the detection of any landmarks, and large intra-class diversities may occur. To address this, we further propose a palmprint feature alignment and learning hybrid network (PalmALNet) for unconstrained palmprint recognition. Specifically, we first exploit and align the multi-scale shallow representation of unconstrained palmprint images via deformable convolution and alignment-aware supervision, such that the pixel gaps of the intra-class palmprint CROIs can be minimized in shallow feature space. Then, we develop multiple triple-attention learning modules by integrating spatial, channel, and self-attention operations into convolution to adaptively learn and highlight the latent identity-invariant palmprint information, enhancing the overall discriminative power of the palmprint features. Extensive experimental results on four challenging palmprint databases demonstrate the promising effectiveness of both the proposed PalmALNet and CROI for unconstrained palmprint recognition.
Chapter
Hand segmentation, as a key part of human-computer interaction, palmprint recognition and gesture recognition, is prone to interference from complex backgrounds subsequently resulting in poor segmentation accuracy. In this paper, a paired spatial U-Net (PSU-Net) hand image segmentation network is proposed. Firstly, we improve the traditional dilated pyramid pooling into the paired spatial pyramid pooling (PSPP) module. Through low-dimensional feature pairing, the PSPP can exploit the low-dimensional feature information, thus enhancing the network’s ability to capture edge detail information. Then we design the global attention fusion module (GAF), which can efficiently combine low-dimensional spatial details and high-dimensional semantic information to solve blurred edges in complex backgrounds. Some experimental results on HOF, GTEA and Egohands databases show the proposed approach has quite good performance. The mIOU of PSU-Net can achieve 76.19% on HOF dataset, while the mIOU of DeeplabV3 is 74.45%.KeywordsHand image segmentationAttention mechanismPyramid pooling structure
Chapter
Biometric systems based on hand traits captured using touchless acquisition procedures are increasingly being used for the automatic recognition of individuals due to their favorable trade-off between accuracy and acceptability by users. Among hand traits, palmprint and fingerprints are the most studied modalities because they offer higher recognition accuracy than other hand-based traits such as finger texture, knuckle prints, or hand geometry. For capturing palmprints and fingerprints, touchless and less-constrained acquisition procedures have the advantage of mitigating the problems caused by latent prints, dirty sensors, and skin distortions. However, touchless acquisition systems for palmprints and fingerprints face several challenges caused by the need to capture the hand while it is moving and under varying illumination conditions. Moreover, images captured using touchless acquisition procedures tend to exhibit complex backgrounds, nonuniform reflections, and perspective distortions. Recently, methods such as adaptive filtering, three-dimensional reconstruction, local texture descriptors, and deep learning have been proposed to compensate for the nonidealities of touchless acquisition procedures, thereby increasing the recognition accuracy while maintaining high usability. This chapter presents an overview of the various methods reported in the literature for touchless palmprint and fingerprint recognition, describing the corresponding acquisition methodologies and processing methods.KeywordsBiometricsTouchlessPalmprintFingerprint
Article
Recently, demand for biometric access controls and online payments in smartphones increased, necessitating further investigation and development in this area. This paper proposes a new low-cost palm vein recognition system for smartphones using RGB images. First, we detect and enhance palm vein patterns, using the saturation channel instead of the red channel as in the existing approaches. Then, to address the challenging contactless capturing problems of smartphones such as scale variants, rotation, closed fingers, or rings on hand, we introduce an improved method for the region of interest extraction, based on the convex hull, with a new idea for key vector use. We also designed a new lightweight deep learning-based model for smartphones, which was overlooked in previous palm vein recognition studies. The proposed model comprises suitable blocks of convolution, depthwise separable convolution, inverted residual bottleneck, and spatial pyramid pooling module; in addition, the accuracy is enhanced with fusion strategy. Results show that the proposed model is both smaller and more accurate than related models. The integrated proposed model obtains the best equal error rate, 0.49%, and an inference time of 8 ms.
Article
Full-text available
Most current face recognition systems are designed for indoor, cooperative-user applications. However, even in thus-constrained applications, most existing systems, academic and commercial, are compromised in accuracy by changes in environmental illumination. In this paper, we present a novel solution for illumination invariant face recognition for indoor, cooperative-user applications. First, we present an active near infrared (NIR) imaging system that is able to produce face images of good condition regardless of visible lights in the environment. Second, we show that the resulting face images encode intrinsic information of the face, subject only to a monotonic transform in the gray tone; based on this, we use local binary pattern (LBP) features to compensate for the monotonic transform, thus deriving an illumination invariant face representation. Then, we present methods for face recognition using NIR images; statistical learning algorithms are used to extract most discriminative features from a large pool of invariant LBP features and construct a highly accurate face matching engine. Finally, we present a system that is able to achieve accurate and fast face recognition in practice, in which a method is provided to deal with specular reflections of active NIR lights on eyeglasses, a critical issue in active NIR image-based face recognition. Extensive, comparative results are provided to evaluate the imaging hardware, the face and eye detection algorithms, and the face recognition algorithms and systems, with respect to various factors, including illumination, eyeglasses, time lapse, and ethnic groups.
Conference Paper
Full-text available
Palmprint-based personal identification, as a new member in the biometrics family, has become an active research topic in recent years. Although great progress has been made, how to represent palmprint for effective classification is still an open problem. In this paper, we present a novel palmprint representation - ordinal measure, which unifies several major existing palmprint algorithms into a general framework. In this framework, a novel palmprint representation method, namely orthogonal line ordinal features, is proposed. The basic idea of this method is to qualitatively compare two elongated, line-like image regions, which are orthogonal in orientation and generate one bit feature code. A palmprint pattern is represented by thousands of ordinal feature codes. In contrast to the state-of-the-art algorithm reported in the literature, our method achieves higher accuracy, with the equal error rate reduced by 42% for a difficult set, while the complexity of feature extraction is halved.
Conference Paper
Full-text available
The ability to detect a persons unconstrained hand in a natural video sequence has applications in sign language, gesture recognition and HCl. This paper presents a novel, unsupervised approach to training an efficient and robust detector which is capable of not only detecting the presence of human hands within an image but classifying the hand shape. A database of images is first clustered using a k-method clustering algorithm with a distance metric based upon shape context. From this, a tree structure of boosted cascades is constructed. The head of the tree provides a general hand detector while the individual branches of the tree classify a valid shape as belong to one of the predetermined clusters exemplified by an indicative hand shape. Preliminary experiments carried out showed that the approach boasts a promising 99.8% success rate on hand detection and 97.4% success at classification. Although we demonstrate the approach within the domain of hand shape it is equally applicable to other problems where both detection and classification are required for objects that display high variability in appearance.
Article
Recently, biometric palmprint has received wide attention from researchers. It is well-known for several advantages such as stable line features, low-resolution imaging, low-cost capturing device, and user-friendly. In this paper, an automated scanner-based palmprint recognition system is proposed. The system automatically captures and aligns the palmprint images for further processing. Several linear subspace projection techniques have been tested and compared. In specific, we focus on principal component analysis (PCA), fisher discriminant analysis (FDA) and independent component analysis (ICA). In order to analyze the palmprint images in multi-resolution-multi-frequency representation, wavelet transformation is also adopted. The images are decomposed into different frequency subbands and the best performing subband is selected for further processing. Experimental result shows that application of FDA on wavelet subband is able to yield both FAR and FRR as low as 1.356 and 1.492% using our palmprint database.
Article
vi, 108 p. : ill. ; 30 cm. PolyU Library Call No.: [THS] LG51 .H577M COMP 2002 Kong Using texture analysis as a tool to extract biometric features for personal identification is the main goal of this thesis. In this study, three biometrics, the iris, paimprint and ear are investigated. Each of them shows a different level of achievement. As far as iris recognition is concerned, we propose a new noise detection model for accurate segmentation of an iris. Eyelashes, the eyelids and reflection are the three main sources of noise. The eyelid issue has been solved by the traditional eye model; however, eyelashes and reflection have yet to be addressed. To determinate a pixel belonging to an eyelash, our model follows three criteria: 1) separable eyelash condition, 2) non-informative condition and 3) connective criterion. For reflection, strong reflection points are detected by a threshold and the weak reflection areas around the strong points are determined by a connective criterion and a statistical test. Using Boles's [47-49] texture-based iris recognition approach to evaluate the accuracy and usefulness of our detection model, we find the experimental results encouraging. For palmprint identification, we develop a novel textured feature extraction technique, in which a 2-D Gabor filter is used to obtain the texture information and two palmprint images are compared by their hamming distance. The experiments give impressive results and show that our method is effective and comparable with fingerprint (FingerCode), iris (IrisCode) and 3-D hand geometry. For ear recognition, we consider two issues: 1) image acquisition and 2) textured feature extraction technique. We have developed a special device for image acquisition. We also propose a novel feature extraction for ear recognition that measures two ear features by a simple vector norm. The experimental results show that ear recognition can provide middle level security. M.Phil., Dept. of Computing, The Hong Kong Polytechnic University, 2002.
Article
Biometrics-based personal identification is regarded as an effective method for automatically recognizing, with a high confidence, a person's identity. This paper presents a new biometric approach to online personal identification using palmprint technology. In contrast to the existing methods, our online palmprint identification system employs low-resolution palmprint images to achieve effective personal identification. The system consists of two parts: a novel device for online palmprint image acquisition and an efficient algorithm for fast palmprint recognition. A robust image coordinate system is defined to facilitate image alignment for feature extraction. In addition, a 2D Gabor phase encoding scheme is proposed for palmprint feature extraction and representation. The experimental results demonstrate the feasibility of the proposed system. Author name used in this publication: Wai-Kin Kong