4
MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE CONTACT-FREE PALMPRINT RECOGNITION Ying Hao, Zhenan Sun, Tieniu Tan and Chao Ren National Laboratory of Pattern Recognition, Institute of Automation, CAS P. O. Box 2728, Beijing, China, 100080 ABSTRACT In this paper, we propose to improve the verification performance of contract-free palmprint recognition system by means of feature-level image registration and pixel-level fusion of multi-spectral palm im- ages. Our method involves image acquisition via a dedicated device under contact-free and multi-spectral environment, preprocessing to locate Region of Interest (ROI) from each individual hand images, feature-level registration to align ROIs from different spectral im- ages in one sequence and fusion to combine images from multiple spectra. The advantages of the proposed method include better hy- giene and higher verification performance. Given a database com- posed of images from 330 hands, two out of four state of the art fu- sion strategies offer significant performance gain and the best Equal Error Rate (EER) is up to 0.5%. Index Termsmulti-spectral imaging, hand biometrics, pixel level fusion, feature level registration, ordinal measure 1. INTRODUCTION In everyday life, hand serves as the most natural tool for human be- ing to perceive and reconstruct surrounding environments. There- fore, its prevalence in the field of Biometrics is no surprising. Al- though higher and higher recognition rates are reported in the liter- ature, the success of most hand based modalities relies on a contact device with pegs for image acquisition, which may invoke hygiene concern and reluctance of use. Recently, a growing trend towards relieving the users from contact device has emerged and the idea of peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro- posed. However, the accuracy of hand biometrics system degrades along with the removal of peg or contact plate due to larger degree of freedom of hands. Combining multiple imaging modalities for better performance has been proved to be feasible in face recognition[4, 5]. In the field of hand biometrics, Wang and Leedham[6] demonstrated in their pi- oneering work that no observable meaningful information of vein pattern can be obtained by passively imaging the infrared radiation emitted by palm side skin. Hence we focus on the electromagnetic spectrum from visible light to near infrared, design a dedicated hard- ware for multi-spectral image acquisition and attempt to combine the correlative and complementary information of multi-spectral im- ages by pixel level fusion. Recently, Wang et al.[7] proposed to fuse palmprint and palm vein images with fusion applied in image level. Our work differs from theirs in the following two aspects: 1. Instead This work is supported by the National Basic Research Program of China (Grant No. 2004CB318100), the National Natural Science Foundation of China (Grant No. 60736018, 60335010, 60702024), the National Hi-Tech Research and Development Program of China (Grant No.2006AA01Z193, 2007AA01Z162), and the Chinese Academy of Sciences. of adopting a frequency-division fashion in image acquisition where more than one camera is required, we choose a time-division strat- egy. Only one camera is utilized and different spectrum images are obtained during sequential time slots, which provides better scala- bility and higher performance price ratio. 2. Compared to the semi touchless image capture scheme used in [7], our device works under a fully contact free environment. In consequence, different prepro- cessing and registration methods are developed. In our previous work[8], we exploited four image fusion strate- gies and compared their performances in sense of root mean square error(RMSE), mutual information(MI), universal image quality in- dex(UIQI) and local average entropy gain between feature level rep- resentations of palmprint and fused images. In this paper, our work was extended to meet updated hardware and enlarged database. The proposed method is organized in Section 2, followed by Section 3 describing the database and experiments that are applied to validate the proposed algorithm. Finally, conclusion and discussion are pre- sented in Section 4. 2. PROPOSED METHOD The proposed method is composed of four steps, as demonstrated in Fig.1. A sequence of multi-spectral hand images is first obtained by illuminating the hand with multiple active lights. Afterwards, preprocessing is conducted independently on each image to achieve coarse localization of ROIs. The localization results in each se- quence of images are then further refined through feature level image registration. Finally, fused image integrating multiple image sources is produced and fed into the verification engine for performance eval- uation. Fig. 1. Flowchart of the proposed method 2.1. Hardware setup Fig.2 illustrates the design of our self-designed image acquisition device. During imaging process, six groups of LEDs ranging from violet to near infrared are turned on sequentially, resulting in a se- quence composed of six hand images. These images are captured in a reflective way under the sheltered environment as shown in the fig- ure. To provide even illumination, each group of lights is arranged on the board in a circular way and scattered by a ground glass be- fore reaching the hand. Depending on wavelength of incident light

MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE …nlpr-web.ia.ac.cn/2008papers/gjhy/gh121.pdf · of peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro-posed. However,

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE …nlpr-web.ia.ac.cn/2008papers/gjhy/gh121.pdf · of peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro-posed. However,

MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE CONTACT-FREE PALMPRINTRECOGNITION

Ying Hao, Zhenan Sun, Tieniu Tan and Chao Ren

National Laboratory of Pattern Recognition, Institute of Automation, CASP. O. Box 2728, Beijing, China, 100080

ABSTRACTIn this paper, we propose to improve the verification performance ofcontract-free palmprint recognition system by means of feature-levelimage registration and pixel-level fusion of multi-spectral palm im-ages. Our method involves image acquisition via a dedicated deviceunder contact-free and multi-spectral environment, preprocessing tolocate Region of Interest (ROI) from each individual hand images,feature-level registration to align ROIs from different spectral im-ages in one sequence and fusion to combine images from multiplespectra. The advantages of the proposed method include better hy-giene and higher verification performance. Given a database com-posed of images from 330 hands, two out of four state of the art fu-sion strategies offer significant performance gain and the best EqualError Rate (EER) is up to 0.5%.

Index Terms— multi-spectral imaging, hand biometrics, pixellevel fusion, feature level registration, ordinal measure

1. INTRODUCTION

In everyday life, hand serves as the most natural tool for human be-ing to perceive and reconstruct surrounding environments. There-fore, its prevalence in the field of Biometrics is no surprising. Al-though higher and higher recognition rates are reported in the liter-ature, the success of most hand based modalities relies on a contactdevice with pegs for image acquisition, which may invoke hygieneconcern and reluctance of use. Recently, a growing trend towardsrelieving the users from contact device has emerged and the ideaof peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro-posed. However, the accuracy of hand biometrics system degradesalong with the removal of peg or contact plate due to larger degreeof freedom of hands.

Combining multiple imaging modalities for better performancehas been proved to be feasible in face recognition[4, 5]. In the fieldof hand biometrics, Wang and Leedham[6] demonstrated in their pi-oneering work that no observable meaningful information of veinpattern can be obtained by passively imaging the infrared radiationemitted by palm side skin. Hence we focus on the electromagneticspectrum from visible light to near infrared, design a dedicated hard-ware for multi-spectral image acquisition and attempt to combinethe correlative and complementary information of multi-spectral im-ages by pixel level fusion. Recently, Wang et al.[7] proposed to fusepalmprint and palm vein images with fusion applied in image level.Our work differs from theirs in the following two aspects: 1. Instead

This work is supported by the National Basic Research Program of China(Grant No. 2004CB318100), the National Natural Science Foundation ofChina (Grant No. 60736018, 60335010, 60702024), the National Hi-TechResearch and Development Program of China (Grant No.2006AA01Z193,2007AA01Z162), and the Chinese Academy of Sciences.

of adopting a frequency-division fashion in image acquisition wheremore than one camera is required, we choose a time-division strat-egy. Only one camera is utilized and different spectrum images areobtained during sequential time slots, which provides better scala-bility and higher performance price ratio. 2. Compared to the semitouchless image capture scheme used in [7], our device works undera fully contact free environment. In consequence, different prepro-cessing and registration methods are developed.

In our previous work[8], we exploited four image fusion strate-gies and compared their performances in sense of root mean squareerror(RMSE), mutual information(MI), universal image quality in-dex(UIQI) and local average entropy gain between feature level rep-resentations of palmprint and fused images. In this paper, our workwas extended to meet updated hardware and enlarged database. Theproposed method is organized in Section 2, followed by Section 3describing the database and experiments that are applied to validatethe proposed algorithm. Finally, conclusion and discussion are pre-sented in Section 4.

2. PROPOSED METHOD

The proposed method is composed of four steps, as demonstratedin Fig.1. A sequence of multi-spectral hand images is first obtainedby illuminating the hand with multiple active lights. Afterwards,preprocessing is conducted independently on each image to achievecoarse localization of ROIs. The localization results in each se-quence of images are then further refined through feature level imageregistration. Finally, fused image integrating multiple image sourcesis produced and fed into the verification engine for performance eval-uation.

Fig. 1. Flowchart of the proposed method

2.1. Hardware setup

Fig.2 illustrates the design of our self-designed image acquisitiondevice. During imaging process, six groups of LEDs ranging fromviolet to near infrared are turned on sequentially, resulting in a se-quence composed of six hand images. These images are captured ina reflective way under the sheltered environment as shown in the fig-ure. To provide even illumination, each group of lights is arrangedon the board in a circular way and scattered by a ground glass be-fore reaching the hand. Depending on wavelength of incident light

Page 2: MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE …nlpr-web.ia.ac.cn/2008papers/gjhy/gh121.pdf · of peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro-posed. However,

and skin condition, variable amount of light is reflected at the skinsurface and the reminder enters the skin to different layers, either ab-sorbed or scattered, revealing internal skin structure to imaging de-vice. Instead of touching any tangible plate, the subjects only needto naturally stretch their hands, with their palms facing the camera.Such a scheme features better hygiene and helps to reduce potentialresistance of use.

Fig. 2. Self-designed imaging device for multi-spectral hand imageacquisition

2.2. Preprocessing

Better human computer interface usually comes together with morecomputation. In our case, the removal of localization plate intro-duces more degree of freedom in hand images. In-depth rotation,pose variation and scale change are most frequently encounteredones among others. Since only the central region of palm is usedfor feature extraction, the goal of preprocessing is to robustly locateROI in presence in above mentioned noises.

Fig. 3. Scale-invariant ROI localization regulation, where size lengthof ROI d is estimated in proportion to palm width

The proposed ROI localization regulation is demonstrated in Fig.3.Compared with traditional palmprint preprocessing method aimingat segmenting images obtained with contact device, the main differ-ence of our method lies in that the size of ROI is estimated rather thanfixed. To do this, we examine a parallel strip 30 - 80 pixels beneaththe reference line and estimate the average palm width, denoted byPW . The side length d of ROI is then set to be d = 0.6 × PW .Afterwards, a coordinate system is established, where origin O is the

middle point of reference line segment minus an offset proportionalto PW and the direction of horizontal axis is defined to be orthog-onal to reference line. Finally, a square region with side length d ispicked out as ROI. To make the localization result scale invariant,the distance from O to ROI is also proportional to PW .

2.3. Feature-level fine registration

To perform pixel level fusion, accurate alignment of the source im-ages is essential. However, due to hand/finger vibration and imag-ing quality variation across wavelengths, ROIs of multi-spectral im-ages are sometimes not precisely aligned with each other. Fig.4(a)demonstrates such an example by applying the above mentioned pre-processing component independently to each image in one sequence.During image sequence acquisition, the subjects were instructed notto move their hands, therefore the assumption that there is no in-depth rotation and scale variation across wavelengths is made. Withthe assumption being approximately valid in the context of our sce-nario, the search space of registration Θ is accordingly reduced torigid transformation and is empirically defined as follows:

Θ = {x, y, θ|10 ≤ x ≤ 10, 10 ≤ y ≤ 10,−3 ≤ θ ≤ 3}where x, yand θ means horizontal and vertical translation and rota-tion angle respectively.

In the field of remote sensing and medical image analysis, mu-tual information (MI) is frequently reported to be effective and ro-bust in tackling registration problem. However, mutual informationonly works well when there exist significant geometrical structures.Although principal lines of palmprint serve this purpose well, theytake only a considerable small portion of image pixels, which re-markably restricts their contribution to registration. Fig.4(c) is anexample of the noisy registration surface based on mutual informa-tion of intensities with regard to the first two registration parametersx and y.

To tackle this problem, we turned to feature level representa-tion other than original intensity of images. Binary textual features,such as CompetitiveCode and ordinal filter [9, 10], have already beenproved to be effective in palmprint recognition task. These methodsseek to robustly, qualitatively and densely represent line-like pat-terns, such as wrinkles, principal lines and ridges. Fig.4(b) illustratesthe Orthogonal Line Ordinal Feature (OLOF)[10] representations ofFig.4(a). Unlike intensity representation, a large amount of consis-tent structures emerge in feature level representation, which are moresuitable for mutual information based registration. Fig.4(d) showsthe new registration curve. It contains a single peak at the expectedposition (12, 12) and is much more smooth than Fig.4(c) elsewhere.

Suppose IWL1, IWL2, · · · , IWLN denote images in one sequence,where N is the total number of wavelengths. The proposed regis-tration is conducted in a sequential way, which means that IWL2

is first registered with regard to IWL1, then IWL3 is registered toIWL2 and so on. Since only images from nearby wavelengths areregistered with each other, registration uncertainty originating fromabrupt changes, which usually occur when active light turns fromvisible to infrared, is circumvented. Fig.4(e) presents the final reg-istration result of (a) and the alignment errors in (a) are successfullycorrected even though the visual appearances changed significantlyfrom IWL1 to IWL6.

2.4. Fusion strategy

According to the generic framework proposed by Zhang et al.[11],an image fusion scheme is usually composed of (a) multiscale de-composition, which maps source intensity images to more efficient

Page 3: MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE …nlpr-web.ia.ac.cn/2008papers/gjhy/gh121.pdf · of peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro-posed. However,

(a)

(b)

05

1015

20

0

10

201.118

1.12

1.122

1.124

1.126

1.128

Parameter 1Parameter 2

Mut

ual I

nfor

mat

ion

(c)

05

1015

20

0

10

201.46

1.48

1.5

1.52

Parameter 1Parameter 2

Mut

ual I

nfor

mat

ion

(d)

(e)

Fig. 4. Feature-level mutual information based registration. (a) pre-processing results of one image sequence, which are regarded ascoarse localization; (b) the three rows denote OLOF representationsof (a) produced by 0◦, 30◦ and 60◦ filters respectively; (c) registra-tion surface based on intensity images; (d) registration surface basedon OLOF representations; (e)ROIs after feature-level registration.

representation; (b) activity measurement that determines the qualityof each input; (c) coefficient grouping method to determine whetheror not cross scale correlation is considered; (d) coefficient combin-ing method where a weighted sum of source representations are cal-culated and finally (e) consistency verification to ensure neighboringcoefficients are calculated in similar manner. Following our previouswork[8], we utilize the following strategy:

• Multiscale decomposition- Based on the observation that line-like pattern are most discriminative features in both visibleand infrared hand images (principal lines and wrinkles forvisible image and blood vessel for infrared image), four mul-tiscale decomposition methods that are frequently referred inthe literatures as being suitable for preserving directional in-formation are chosen. More precisely, they are gradient pyra-mid(GP), morphological pyramid(MP), shift-invariant digitalwavelet transform(SIDWT) and curvelet transform(CT).

• Activity measure- The absolute value of each coefficient istaken as activity measure of each scale, position and some-times orientation.

• Coefficient combining method- For simplicity, we ignoredcross scale correlation and target coefficients in high frequencyare selected by choosing the maximum among source multi-

Table 1. Comparison of verification performance of each wave-length

wavelength index EER d′

WL1 0.67% 6.0625WL2 0.72% 5.8052WL3 0.70% 5.9127WL4 0.90% 5.5068WL5 0.90% 5.9249WL6 0.72% 6.1920

scale representations while for base approximation, weightedsum of source images is utilized.

3. EXPERIMENTAL RESULTS

Experiments are conducted on a multi-spectral database recently col-lected in CASIA. 165 volunteers ranging from 20 to 63 years oldparticipate in the database setup. For each person, we take 3 imagesequences of both hands. Five wavelengths from violet to near in-frared (400 - 1000nm) as well as white light are selected for eachsequence, thus the final database contains 165× 2× 3× 6 = 5940images. Each original image is of size 768 × 576, stored as 8-bitgrayscale image. The resulting side length of ROI d approximatelyfollows normal distribution, with mean value µ = 166 and standarddeviation σ = 16. The ROIs are normalized to 128 × 128 afterregistration.

120 130 140 150 160 170 180 190 200 210 220 2300

20

40

60

80

100

120

140

160

180

ROI side length: d

Empirical distribution of dGaussian fitting of empirical distribution

Fig. 5. Distribution of ROI side length

OLOF feature, as described in [10], is selected to represent ROIsand fused image. Hamming Distance is used to measure the dissim-ilarity between two features. Table.1 compared the verification per-formance of different wavelengths via two important performanceindices: Equal Error Rate(EER) and discriminating index(d′). Ex-cept for WL4 and WL5, the remaining four wavelengths demon-strates comparable recognition potentials. The poor performancesof WL4 and WL5 may be explained by the fact that they bothfall in the region between visible and infrared light and the conse-quence is neither superficial nor vascular information is well pre-served. Although reflecting different visual contents, the visiblespectrum(WL1, WL2 and WL3) and infrared light(WL6) producesimilar EERs, which motivate us to fuse them for better verificationaccuracy.

Image fusion is performed by combining two selected wave-lengths and two combinations, namely (WL1, WL6) and(WL3, WL6) are tested since they are considered to convey com-plementary information. To give a visual impression, Fig.6 showsimages with columns corresponding to the four multi-scale decom-position methods and rows denoting the two combination respec-tively. Although morphological pyramid based method seeminglybest preserves vein pattern, it sometimes introduces blur while curvelettransform constantly maintains sharp line singularities even on fusedimages.

The two combinations are also tested and evaluated in sense

Page 4: MULTISPECTRAL PALM IMAGE FUSION FOR ACCURATE …nlpr-web.ia.ac.cn/2008papers/gjhy/gh121.pdf · of peg-free, or further contact-free, hand biometrics[1, 2, 3] is pro-posed. However,

(a)

(b)

Fig. 6. Fused images using proposed methods, (a) WL1 and WL6fused image using GP, MP, SIDWT and CT from left to right respec-tively; (b) WL3 and WL6 fused image using GP, MP, SIDWT andCT respectively.

Table 2. Verification performance of fused imagesImage WL1+WL6 WL3+WL6

fusion methods EER d′ EER d′

GP 0.71% 5.9215 0.61% 5.8647MP 0.91% 5.5339 0.74% 5.5171

SIDWT 0.50% 6.0441 0.61% 5.9513CT 0.50% 6.1804 0.58% 5.9608

of verification performance, which is listed in Table 2. Obviously,curvelet transform, shift-invariant digital Wavelet transform constantlyhelp to improve verification performance, with curvelet transformperforming best. The excellent capability of curvelet transform inpreserving curve singularities contributes most to performance gain.

4. CONCLUSION AND DISCUSSION

The growing trend of intelligent human computer interface has mo-tivated us to tackle the problem of contact free hand based biomet-rics. Human skin, complex as it is, provides much more informationbeyond the outmost appearance. In this work, we took a first steptowards integrating information deep inside skin with appearance inthe context of hand biometrics. Our contribution is summarized inthe following:

• We have demonstrated that it is feasible to improve perfor-mance of contact free hand based biometrics via pixel-levelfusion of multispectral images. The advantages of pixel-levelfusion are two folds. On the one hand, appearance as wellas inner information of hand is combined to form one solorepresentation, enforcing the security of the whole system.Generally, forged sample can only imitate only one aspect ofskin, which is less likely to be accepted by the system regis-tered with fused sample; On the other hand, this scheme doesnot introduce extra memory consumption, which makes fastrecognition possible once the number of users is large.

• Despite the widely use of mutual information in image regis-tration, it does not produce satisfying results in our scenariodue to lack of structural information. A feature-level registra-tion is proposed instead, providing much better alignment.

• It is normally believed that vein pattern is essentially sparse,which means its discriminative information takes form of iso-lated point or line feature(corresponding to ending, crossoverof blood vessel and vein). Our work demonstrates that dense

representation can also provide satisfying performance. Theunderlying reason lies in the fact that images captured underinfrared spectrum contain more information than pure bloodvessels. Some of the superficial information, such as princi-pal lines, is also visible.

In the future, we plan to explore the possibility of hierarchicalfusion which takes full advantages of the imaging capability of ourdevice. We believe that even better performance can be achieved byintelligently fusing more information. For example, principal linesusually have smaller widths if compared with blood vessels. If dif-ferent scale ordinal measure features are extracted from the samefused image, emphasizing the principal lines and veins respectively,the overall performance might be further improved.

5. REFERENCES

[1] Wei Xiong, Changsheng Xu, and Sim Heng Ong, “Peg-freehuman hand shape analysis and recognition,” in Proceedings ofInt. Conf. on Acoustics, Speech, and Signal Processing, 2005,pp. 77–80.

[2] Xiaoqian Jiang, Wanhong Xu, Latanya Sweeney, Yiheng Li,Ralph Gross, and Daniel Yurovsky, “New directions in contactfree hand recognition,” in Proceedings of Int. Conf. on ImageProcessing, 2007, pp. 389–392.

[3] Wei Xiong, Kar-Ann Toh, Wei-Yun Yau, and Xudong Jiang,“Model-guided deformable hand shape recognition withoutpositioning aids,” Pattern Recognition, vol. 38, 2005.

[4] Seong G. Kong, Jingu Heo, Faysal Boughorbel, Yue Zheng,Besma R. Abidi Andreas Koschan, Mingzhong Yi, andMongi A. Abidi, “Multiscale fusion of visible and thermal irimages for illumination-invariant face recognition,” Int. Jour-nal of Computer Vision, vol. 71, 2007.

[5] Richa Singh, Mayank Vatsa, and Afzel Noore, “Integratedmultilevel image fusion and match score fusion of visible andinfrared face images for robust face recognition,” Int. Journalof Computer Vision, vol. 71, 2007.

[6] Lingyu Wang and Graham Leedham, “Near- and far- infraredimaging for vein pattern biometrics,” in Proc. of the IEEE Intl.Conf. on Video and Signal Based Surveillance, 2006.

[7] Jian-Gang Wang, Wei-Yun Wang, Andy Suwandy, and EricSung, “Fusion of palmprint and palm vein images for personrecognition based on ’laplacianpalm’ feature,” in IEEE Com-puter Vision and Pattern Recognition Workshop on Biometrics,2007, pp. 1–8.

[8] Ying Hao, Zhenan Sun, and Tieniu Tan, “Comparative studieson multispectral palm image fusion for biometrics,” in LectureNotes in Computer Science - ACCV 2007.

[9] Adams Wai-Kin Kong and David Zhang, “Competitive codingscheme for palmprint verification,” in Proceedings of Int. Conf.on Pattern Recognition. IEEE, 2004, pp. 520– 523.

[10] Zhenan Sun, Tieniu Tan, Yunhong Wang, and Stan Z. Li, “Or-dinal palmprint recognition for personal identification,” in Pro-ceedings of Conf. on Computer Vision and Pattern Recognition,2005.

[11] Zhong Zhang and Rick S. Blum, “A categorization ofmultiscale-decomposition-based image fusion schemes with aperformance study for a digital camera application,” in Proc.of IEEE.