IRIS RECOGNITION BASED ON HILBERT–HUANG TRANSFORM 1 ...
IRIS RECOGNITION BASED ON HILBERT–HUANG TRANSFORM 1 ...
IRIS RECOGNITION BASED ON HILBERT–HUANG TRANSFORM 1 ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Advances in Adaptive Data Analysis<br />
Vol. 1, No. 4 (2009) 623–641<br />
c○ World Scientific Publishing Company<br />
<strong>IRIS</strong> <strong>RECOGNITI<strong>ON</strong></strong> <strong>BASED</strong> <strong>ON</strong><br />
<strong>HILBERT–HUANG</strong> <strong>TRANSFORM</strong><br />
ZHIJING YANG∗ , ZHIHUA YANG † and LIHUA YANG∗,‡ ∗School of Mathematics and Computing Science<br />
Sun Yat-sen University, Guangzhou 510275, China<br />
† Information Science School<br />
GuangDong University of Business Studies<br />
Guangzhou 510320, China<br />
‡ mcsylh@mail.sysu.edu.cn<br />
As a reliable approach for human identification, iris recognition has received increasing<br />
attention in recent years. This paper proposes a new analysis method for iris recognition<br />
based on Hilbert–Huang transform (HHT). We first divide a normalized iris image into<br />
several subregions. Then the main frequency center information based on HHT of each<br />
subregion is employed to form the feature vector. The proposed iris recognition method<br />
has nice properties, such as translation invariance, scale invariance, rotation invariance,<br />
illumination invariance and robustness to high frequency noise. Moreover, the experimental<br />
results on the CASIA iris database which is the largest publicly available iris<br />
image data sets show that the performance of the proposed method is encouraging and<br />
comparable to the best iris recognition algorithm found in the current literature.<br />
Keywords: Iris recognition; empirical mode decomposition (EMD); Hilbert–Huang transform<br />
(HHT); main frequency center.<br />
1. Introduction<br />
With increasing demands in automated personal identification, biometric authentication<br />
has been receiving extensive attention over the last decade. Biometrics<br />
employs various physiological or behavioral characteristics, such as fingerprints,<br />
face, iris, retina and palmprints, etc., to accurately identify each individual. 13,29<br />
Among these biometric techniques, iris recognition is tested as one of the most<br />
accurate manner of personal identification. 4,6,14,22,23<br />
The human iris, a thin circular diaphragm lying between the cornea and the lens,<br />
has an intricate structure and provides many minute characteristics such as furrows,<br />
freckles, crypts, and coronas. 2 These visible characteristics, which are generally<br />
called the texture of the iris, are unique to each subject. 6,7,14,24 The iris patterns of<br />
the two eyes of an individual or those of identical twins are completely independent<br />
and uncorrelated. Additionally, the iris is highly stable over a person’s lifetime<br />
and lends itself to noninvasive identification since it is an externally visible internal<br />
623
624 Z.Yang,Z.Yang&L.Yang<br />
organ. All these desirable properties make iris recognition suitable for highly reliable<br />
personal identification.<br />
For the last decade, a number of researchers have worked on iris recognition and<br />
have achieved great progress. According to the various feature extractions, existing<br />
iris recognition methods can be roughly divided into four major categories: the<br />
phase-based methods, 6,7,23 the zero-crossing representation-based methods, 4,22 the<br />
texture analysis-based methods 14,24 and intensity variation analysis methods. 15,16<br />
A phase-based method is a process of phase demodulation. Daugman 6,7 made use<br />
of multiscale Gabor filters to demodulate texture phase structure information of<br />
the iris. Then the filter outputs were quantized to generate a 2048-bit iriscode<br />
to describe an iris. Tisse et al. 23 encoded the instantaneous phase and emergent<br />
frequency with the analytic image (two-dimensional Hilbert transform) as iris features.<br />
The zero-crossings of wavelet transform provide meaningful information of<br />
image structures. Boles and Boashash 4 calculated zero-crossing representation of<br />
one-dimensional wavelet transform at various resolution levels of a virtual circle on<br />
an iris image to characterize the texture of the iris. Wildes et al. 24 represented the<br />
iris texture with a Laplacian pyramid constructed with four different resolution levels.<br />
Tan et al. 14 proposed a well-known texture analysis method by capturing both<br />
global and local details from an iris with the Gabor filters at different scales and<br />
orientations. As an intensity variation analysis method, Tan et al. constructed a set<br />
of one-dimensional intensity signals to contain the most important local variations<br />
of the original two-dimensional iris image. Then the Gaussian–Hermite moments of<br />
such intensity signals are used as distinguishing features. 16<br />
Fourier and Wavelet descriptors have been used as powerful tools for feature<br />
extraction which is a crucial processing step for pattern recognition. However, the<br />
main drawback of those methods is that their basis functions are fixed and do not<br />
necessarily match the varying nature of signals. Hilbert–Huang transform (HHT)<br />
developed by Huang et al. is a new analysis method for nonlinear and nonstationary<br />
data. 11 It can adaptively decompose any complicated data set into a finite<br />
number of intrinsic mode functions (IMFs) that become the bases representing the<br />
data by empirical mode decomposition (EMD). With Hilbert transform, the IMFs<br />
yield instantaneous frequencies as functions of time. The final presentation of the<br />
results is a time–frequency–energy distribution, designated as the Hilbert spectrum<br />
that gives sharp identifications of salient information. Therefore, it brings not<br />
only high decomposition efficiency but also sharp frequency and time localizations.<br />
Recently, the HHT has received more attention in terms of interpretations 9,18,21 and<br />
applications. Its applications have spread from ocean science, 10 biomedicine, 12,20<br />
speech signal processing, 25 image processing, 3 pattern recognition 19,26,28 and<br />
so on. Recently, EMD is also used for iris recognition as a low pass filter<br />
in Ref. 5.<br />
Since a random iris pattern can be seen as a texture, many well-developed texture<br />
analysis methods have been adapted to recognize the iris. 14,24 An iris consists<br />
of some basic elements which are similar to each other and interlaced each other,
Iris image preprocessing<br />
Iris image Localization<br />
Normalization<br />
Iris Recognition Based on Hilbert–Huang Transform 625<br />
Feature<br />
extraction Classification<br />
HHT LDA<br />
Fig. 1. Diagram of the proposed method.<br />
Output<br />
i.e. an iris image is generally periodical to some extent. Therefore the approximate<br />
period is an effective feature for the iris recognition. By employing the main frequency<br />
center presented in our previous works 26,27 of the Hilbert marginal spectrum<br />
as an approximation for the period of an iris image, a new iris recognition method<br />
based on HHT is proposed in this paper. Unlike directly using the residue of the<br />
EMD decomposed iris image for recognition in Ref. 5, the proposed method utilizes<br />
the main frequency center information as the feature vector which is particularly<br />
rotation invariant. In comparison with the existing iris recognition methods, the<br />
proposed algorithm has an excellent percentage of correct classification, and possesses<br />
very nice properties, such as translation invariance, scale invariance, rotation<br />
invariance, illumination invariance and robustness to high frequency noise. Figure 1<br />
illustrates the main steps of our method.<br />
The remainder of this paper is organized as follows. Brief descriptions of image<br />
preprocessing are provided in Sec. 2. A new feature extraction method and matching<br />
are given in Sec. 3. Experimental results and discussions are reported in Sec. 4.<br />
Finally, conclusions of this paper are summarized in Sec. 5.<br />
2. Iris Image Preprocessing<br />
An iris image, contains not only the iris but also some irrelevant parts (e.g. eyelid,<br />
pupil, etc.). A change in the camera-to-eye distance may also result in variations in<br />
the size of the same iris. Therefore, before feature extraction, an iris image needs<br />
to be preprocessed to localize and normalize. Since a full description of the preprocessing<br />
method is beyond the scope of this paper, such preprocessing is introduced<br />
briefly as follows.<br />
The iris is an annular part between the pupil (inner boundary) and the sclera<br />
(outer boundary). Both the inner boundary and the outer boundary of a typical iris<br />
can approximately be taken as circles. This step detects the inner boundary and<br />
the outer boundary of the iris. Since the localization method proposed in Ref. 14 is<br />
a very effective method, we adopt it here. The main steps are briefly introduced as<br />
follows. Since the pupil is generally darker than its surroundings and its boundary<br />
is a distinct edge feature, it can be found by using edge detection (Canny operator<br />
in experiments). Then a Hough transform is used to find the center and radius of<br />
the pupil. Finally, the outer boundary will be detected by using edge detection and
626 Z.Yang,Z.Yang&L.Yang<br />
(a) (b)<br />
(c)<br />
Fig. 2. Iris image preprocessing: (a) original image; (b) localized image; (c) normalized image.<br />
Hough transform again in a certain region determined by the center of the pupil.<br />
A localized image is shown in Fig. 2(b).<br />
Irises from different people may be captured in different sizes and, even for<br />
irises from the same eye, the size may change due to illumination variations and<br />
other factors. It is necessary to compensate for the iris deformation to achieve more<br />
accurate recognition results. Here, we counterclockwise unwrap the iris ring to a<br />
rectangular block with a fixed size (64 × 512 in our experiments). 6,14 That is, the<br />
original iris in a Cartesian coordinate system is projected into a doubly dimensionless<br />
pseudopolar coordinate system. The normalization not only reduces to a<br />
certain extent distortion caused by pupil movement but also simplifies subsequent<br />
processing. A normalized image is shown in Fig. 2(c).<br />
3. Feature Extraction and Matching<br />
3.1. The Hilbert–Huang Transform<br />
The Hilbert–Huang Transform (HHT) was proposed by Huang et al., 11 which is<br />
an important method for signal processing. It consists of two parts: the empirical<br />
mode decomposition (EMD) andtheHilbert spectrum. With EMD, any complicated<br />
data set can be decomposed into a finite and often small number of intrinsic mode<br />
functions (IMFs). An IMF is defined as a function satisfying the following two<br />
conditions: (1) it has exactly one zero-crossing between any two consecutive local<br />
extrema; (2) it has zero local mean.
Iris Recognition Based on Hilbert–Huang Transform 627<br />
By the EMD algorithm, any signal x(t) can be decomposed into finite IMFs,<br />
cj(t)(j =1, 2,...,n), and a residue r(t), where n is the number of IMFs, i.e.<br />
x(t) =<br />
n<br />
cj(t)+r(t). (1)<br />
j=1<br />
Having obtained the IMFs by EMD, we can apply the Hilbert transform to each<br />
IMF, cj(t), to produce its analytic signal zj(t) =cj(t) +iH[cj(t)] = aj(t)e iθj(t) .<br />
Therefore, x(t) can also be expressed as<br />
x(t) =Re<br />
n<br />
aj(t)e iθj (t) + r(t). (2)<br />
j=1<br />
Equation (2) enables us to represent the amplitude and the instantaneous frequency<br />
as functions of time in a three-dimensional plot, in which the amplitude is contoured<br />
on the time–frequency plane. The time–frequency distribution of amplitude<br />
is designated as the Hilbert spectrum, denoted by H(f,t) whichgivesatime–<br />
frequency–amplitude distribution of a signal x(t). HHT brings sharp localizations<br />
both in frequency and time domains, so it is very effective for analyzing nonlinear<br />
and nonstationary data.<br />
With the Hilbert spectrum defined, the Hilbert marginal spectrum can be<br />
defined as<br />
h(f) =<br />
T<br />
0<br />
H(f,t)dt. (3)<br />
The Hilbert marginal spectrum offers a measure of total amplitude (or energy)<br />
contribution from each frequency component.<br />
3.2. Main frequency and main frequency center<br />
It is found that the Hilbert marginal spectrum h(f) has some properties, which can<br />
be used to extract features for iris recognition. Specifically, the main frequency center<br />
of the Hilbert marginal spectrum can be served as a feature to identify different<br />
irises. The “main frequency” and “main frequency center” concepts proposed by us<br />
have been clear described and discussed in our previous works. 26,27 We have shown<br />
that the main frequency center can characterize the approximate period very well.<br />
Here, we only review the definitions of main frequency, main frequency center and<br />
other related concepts as follows.<br />
Definition 1 (Main frequency). Let x(t) be an arbitrary time series and h(f)<br />
be its Hilbert marginal spectrum, then fm is called as the main frequency of x(t), if<br />
h(fm) ≥ h(f), ∀f.<br />
Definition 2 (Average Hilbert marginal spectrum of signal series). Let<br />
X = {xj(t)|j =1, 2,...,N}, whereeachxj(t) isatimeseries,andhj(f) bethe
628 Z.Yang,Z.Yang&L.Yang<br />
Hilbert marginal spectrum of xj(t). The average Hilbert marginal spectrum of X<br />
is defined as<br />
H(f) = 1<br />
N<br />
hj(f). (4)<br />
N<br />
j=1<br />
f H m is called as the average main frequency of X if f H m satisfies H(f H m ) ≥ H(f), ∀f.<br />
For a given set of signal series, in which signals are approximately periodic, the<br />
main frequency can characterize the approximate period very well. Unfortunately, in<br />
some cases a signal may not have a unique main frequency. To handle this situation,<br />
all the possible main frequencies have to be considered. Therefore, we can utilize<br />
the gravity frequencies, which is called the “main frequency center”, instead of the<br />
“main frequency”.<br />
Definition 3 (Main frequency center). Let H(fi) (j =1, 2,...,W)bethe<br />
average Hilbert marginal spectrum of X = {xj(t)|j =1, 2,...,N}. Assume H(fi)<br />
is monotone decreasing respect to i. The main frequency center of X is defined as<br />
fC(X) =<br />
M<br />
i=1 fiH(fi)<br />
M<br />
i=1<br />
H(fi) , (5)<br />
where M is the minimum integer satisfying M i=1 H(fi) ≥ P W i=1 H(fi), and 0 <<br />
P < 1 is a given constant. Then e =(1/M ) M i=1 H(fi) is called the energy of<br />
fC(X).<br />
As can be seen from its definition, the main frequency center is the weighted<br />
mean of several frequencies whose Hilbert marginal spectrums are largest. There<br />
and the main frequency center<br />
is a gap between the average main frequency f H m<br />
fC, but the main frequency center will be a steadier recognition feature. Therefore<br />
the main frequency center fC is used instead of the average main frequency f H m to<br />
characterize the approximate period of signal series.<br />
3.3. Iris feature extraction<br />
Although all normalized iris templates have the same size, eyelashes and eyelids<br />
may still appear on the templates and degrade recognition performance. We find<br />
that the upper portion of a normalized iris image (corresponding to regions closer to<br />
the pupil) provides the most useful information for recognition (see Fig. 3). Therefore,<br />
the region of interest (ROI) is selected to remove the influence of eyelashes<br />
and eyelids. That is, the features are extracted only from the upper 75% section<br />
(48 × 512) that is closer to the pupil in our experiments (see Fig. 3). It is known<br />
that an iris has a particularly interesting structure and provides abundant texture<br />
information. Furthermore if we vertically divide the ROI of a normalized image into<br />
three subregions as shown in Fig. 3, the texture information in the subregion will<br />
be more distinct.
ROI {<br />
Iris Recognition Based on Hilbert–Huang Transform 629<br />
Fig. 3. The normalized iris image is vertically divided into four subregions. The three subregions<br />
1,2and3aretheROI.<br />
210<br />
200<br />
190<br />
180<br />
170<br />
160<br />
150<br />
140<br />
130<br />
120<br />
0 50 100 150 200 250 300 350 400 450 500<br />
x<br />
1<br />
2<br />
3<br />
4<br />
200<br />
150<br />
100<br />
0<br />
50<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−50<br />
0<br />
50<br />
50 100 150 200 250 300 350 400 450 500<br />
c1<br />
c2<br />
c3<br />
c4<br />
c5<br />
r<br />
0<br />
−50<br />
0<br />
50<br />
50 100 150 200 250 300 350 400 450 500<br />
0<br />
−50<br />
0<br />
50<br />
50 100 150 200 250 300 350 400 450 500<br />
0<br />
−50<br />
0 50 100 150 200 250 300 350 400 450 500<br />
20<br />
0<br />
−20<br />
0<br />
200<br />
50 100 150 200 250 300 350 400 450 500<br />
150<br />
0 50 100 150 200 250 300 350 400 450 500<br />
Fig. 4. Left, the 11th line signal of the normalized iris image in Fig. 3 along the horizontal<br />
direction. Right, the EMD decomposition result of the 11th line signal.<br />
An iris consists of some basic elements which are similar each other and interlaced<br />
each other. Hence, an iris image is generally periodic to some extent along<br />
some directions, that is, some approximate periods are embed in the iris image. As<br />
an example, let us observe the 11th line signal of the normalized iris image in Fig. 3<br />
along the horizontal direction, as show in the left of Fig. 4. It can be seen that most<br />
of the durations of the waves are similar, i.e. some main frequencies embed in the<br />
signal. As we know, the EMD can extract the low-frequency oscillations very well. 11<br />
With EMD, the decomposition result of the 11th line signal is shown in the right of<br />
Fig. 4. It can be seen that two main approximate periods are extracted in the third<br />
and fourth IMFs. To show it clearly, we plot the original signal (solid line) and the<br />
third IMF (dash line) together in the interval [320, 450] in the left of Fig. 5. It can<br />
be seen that this IMF characterizes the proximate period of the waveform quite well<br />
and the period is about 15 (i.e. the frequency is about 1/15 ≈ 0.067). Similarly, we<br />
plot the original signal (solid line) and the fourth IMF (dash line) together in the<br />
interval [320, 450] in the right of Fig. 5. It can be seen that this IMF characterizes<br />
the proximate period and the variety of the amplitude. This period is about 26 (i.e.<br />
the frequency is about 1/26 ≈ 0.0385).<br />
According to Eq. (3), we can compute the Hilbert marginal spectrum of the<br />
11th line signal of the normalized iris image in Fig. 3, as shown in the left of<br />
Fig. 6. It is evident that the two main frequencies can be extracted from the Hilbert<br />
marginal spectrum correctly. Based on lots of experiments and analysis we found
630 Z.Yang,Z.Yang&L.Yang<br />
210<br />
200<br />
190<br />
180<br />
170<br />
160<br />
150<br />
140<br />
130<br />
120<br />
320 340 360 380 400 ϖϖ<br />
420 440<br />
210<br />
200<br />
190<br />
180<br />
170<br />
160<br />
150<br />
140<br />
130<br />
120<br />
320 340 360 380 400 420 440<br />
Fig. 5. Left, the original signal (solid line) and the third IMF (dash line) in the interval [320,<br />
450]. Right, the original signal (solid line) and the fourth IMF (dash line) in [320, 450].<br />
Frequency Content<br />
700<br />
600<br />
500<br />
400<br />
300<br />
200<br />
100<br />
0<br />
0 0.038 0.068 0.1 0.2<br />
Frequency (Hz)<br />
0.3 0.4 0.45<br />
Frequency Content<br />
700<br />
600<br />
500<br />
400<br />
300<br />
200<br />
100<br />
0<br />
0 0.036 0.069 0.1 0.2<br />
Frequency (Hz)<br />
0.3 0.4 0.45<br />
Frequency Content<br />
700<br />
600<br />
500<br />
400<br />
300<br />
200<br />
100<br />
f C<br />
0<br />
0 0.051 0.1 0.2<br />
Frequency (Hz)<br />
0.3 0.4 0.45<br />
Fig. 6. Left, the Hilbert marginal spectrum of the 11th line signal of the normalized iris in Fig. 3.<br />
Middle, the Hilbert marginal spectrum of the 12th line signal. Right, the average Hilbert marginal<br />
spectrum and the main frequency center fC of the line signal series in the first subregion of the<br />
normalized iris in Fig. 3.<br />
that the main frequency information of signals along the same direction in the same<br />
subregion of an iris is similar. As an example, we compute the Hilbert marginal<br />
spectrum of the 12th line signal of the normalized iris image in Fig. 3, as shown in<br />
the middle of Fig. 6. It can be seen that the Hilbert marginal spectrums of the 12th<br />
signal is very similar with that of the 11th signal. Then we compute the average<br />
Hilbert marginal spectrum of the signal series along the horizontal direction in the<br />
first subregion of Fig. 3, as shown in the right of Fig. 6. It can be seen that it is<br />
not only coincident with the Hilbert marginal spectrum of each line but also more<br />
concentrated. Therefore, based on the average Hilbert marginal spectrum of the<br />
line signal series in each subregion, we can obtain the main frequency center fC<br />
described as Eq. (5) as a reliable feature of the iris.<br />
Since the orientation information is a very important pattern in an iris, 14,24<br />
the main frequency center along the horizontal direction is not only enough to<br />
characterize its texture information. To characterize the orientation information,<br />
the features along the other directions should be considered. First of all, we should
α ( ) α<br />
Iris Recognition Based on Hilbert–Huang Transform 631<br />
Fig. 7. The generation of the signal along angle α of each subregion: choose one point of the first<br />
column and connect all the dotted lines.<br />
present the generation of signal series along angle α of each subregion of an iris.<br />
The generation method can be simply described as follows. As shown in Fig. 7, the<br />
rectangle frame denotes one of the subregion of an iris. Firstly choose one point of<br />
the first column and connect all the dotted lines. Then we obtain one signal along<br />
angle α of the subregion. Similarly, we can generate all the signal series (totally<br />
16 signals) along angle α in the subregion. In experiments, we totally choose 18<br />
directions: 0 ◦ , 10 ◦ ,...,170 ◦ .<br />
It is found that it has a good classification performance when the main frequency<br />
centers described as Eq. (5) are used as features for iris recognition. The features<br />
can cluster the samples of same class of iris. As an example, we show three samples<br />
of the same iris and their main frequency centers of 18 directions in I1 in Fig. 8. It<br />
can be seen that the features cluster very well. Furthermore, the gaps are usually<br />
existent for the samples from different classes of iris images. This implies that the<br />
selected features have really a good classification ability for the iris images.<br />
It is also found that the energy e of the main frequency center is also a good<br />
feature for classification. It can reflect the image contrast of different classes of iris.<br />
The higher the contrast is, the larger energy the image has. In other words, a signal<br />
will have a larger energy if it waves in a larger amplitude. Thus, a larger energy<br />
in marginal spectrum can be expected if an iris has the higher contrast. It can be<br />
seen from Fig. 9 that in these three iris classes (a), (b) and (c), the class (a) has<br />
the highest contrast while the class (c) has the lowest contrast. It indicates that the<br />
class (a) has generally the largest energy while the class (c) has the smallest energy<br />
of the main frequency center along the same orientation. An encouraging result is<br />
received as shown in Fig. 9(d). It implies that the energies of the main frequency<br />
center should be also used as features to classify the irises.<br />
As is known, that we choose the main frequency center and its energy as features,<br />
now the feature extraction algorithm is presented as follows.<br />
Algorithm 1 (Feature extraction algorithm). Given a normalized iris image<br />
I(i, j), let its three subregions from the top down divided as Fig. 3 be I1, I2<br />
and I3.<br />
D = {di|di =(i − 1)10,i=1, 2,...,18} — the feature orientations.<br />
. . . . . .
632 Z.Yang,Z.Yang&L.Yang<br />
Main Frequency Center<br />
0.07<br />
0.06<br />
0.05<br />
0.04<br />
0.03<br />
(a) (b) (c)<br />
0 20 40 60 80<br />
Orientation<br />
100 120 140 160<br />
(d)<br />
Fig. 8. (a)–(c) are three samples of the same iris; (d) their main frequency centers of 18 orientations<br />
in I1.<br />
Step 1: Calculate the main frequency centers and energies of 18 orientations of<br />
I1 to form the main frequency vector, f1, by the following six steps.<br />
(1) Let i =1.<br />
(2) Generate 16 signals along orientation di as described in Fig. 7 to form<br />
the signal set, denoted by X.<br />
(3) Calculate the average Hilbert marginal spectrum of X.<br />
(4) Calculate the main frequency center fC(I1(di)) and its corresponding<br />
energy e(I1(di)) of X.<br />
(5) Let i = i +1, if i ≤ 18, then go back to (2); otherwise go to (6).<br />
(6) Obtain the main frequency vector f1 as follows<br />
f1 =(fC(I1(0)),...,fC(I1(170)),e(I1(0)),...,e(I1(170))).<br />
(a)<br />
(b)<br />
(c)
Energy<br />
600<br />
400<br />
200<br />
Iris Recognition Based on Hilbert–Huang Transform 633<br />
(a) (b) (c)<br />
0 20 40 60 80 100 120 140 160<br />
Orientation<br />
(d)<br />
Fig. 9. (a)–(c) are three samples from three different iris classes; (d) each class contains three<br />
samples. The energies of the main frequency center of 18 orientations in I1.<br />
Step 2: Calculate the main frequency centers and its energies of 18 orientations of<br />
I2 similarly to form the main frequency vector<br />
f2 =(fC(I2(0)),...,fC(I2(170)),e(I2(0)),...,e(I2(170))).<br />
Step 3: Calculate the main frequency centers and its energies of 18 orientations of<br />
I3 similarly to form the main frequency vector<br />
f3 =(fC(I3(0)),...,fC(I3(170)),e(I3(0)),...,e(I3(170))).<br />
Step 4: Finally the feature vector of the iris is defined as<br />
The feature vector contains 108 components.<br />
F =(f1,f2,f3).<br />
(a)<br />
(b)<br />
(c)
634 Z.Yang,Z.Yang&L.Yang<br />
The proposed iris feature vector has nice properties. Let us discuss its invariance<br />
and the robustness to noise as follows.<br />
3.3.1. Invariance<br />
It is desirable to obtain an iris representation invariant to translation, scale, rotation<br />
and illumination. In our method, translation invariance and approximate scale<br />
invariance are achieved by normalizing the original image at the preprocessing step.<br />
Rotation invariance is important for an iris representation since changes of head<br />
orientation and binocular vergence may cause eye rotation. Most existing schemes<br />
achieve approximate rotation invariance either by rotating the feature vector before<br />
matching4,6,7,15,23 or by defining several templates which denote other rotation<br />
angles for each iris class in the database. 5,14 In our algorithm, the annular iris is<br />
unwrapped into a rectangular image. Therefore, rotation in the original image just<br />
corresponds to translation in the normalized image (for example, clockwise rotation<br />
of 90◦ in the original image just corresponds to circle translation 128 pixels towards<br />
the left side in the normalized image, as shown in Figs. 10(a)–(d)). Fortunately,<br />
translation invariance can easily be achieved in our method. Since translation in<br />
the original signal will just result in almost the same translation of all IMFs and<br />
the residue, 11 as shown in Fig. 10(e), (f). Furthermore, if let g(t) =g(t − t0), we<br />
have<br />
H[g](t) = 1<br />
π p.v.<br />
∞<br />
g(t<br />
−∞<br />
′ )<br />
t − t ′ dt′ = 1<br />
π p.v.<br />
∞<br />
g(t<br />
−∞<br />
′ )<br />
(t − t0) − t ′ dt′ = H[g](t − t0).<br />
Therefore, translation in the IMFs just results in the same translation in its Hilbert<br />
transform so as its instantaneous frequencies. Then the Hilbert marginal spectrums<br />
of the original signal and the translation signal will be the same. Finally, the main<br />
frequency center and its energy based on the average Hilbert marginal spectrum of<br />
signal series will also keep the same. Additionally, we can also observe this property<br />
intuitively. Since our feature is based on the vertical subregion of the normalized<br />
image, translation in the normalized image will not change the information of the<br />
subregion. Therefore, the feature will keep the same.<br />
Since the illuminations of iris images may be different caused by the position<br />
of light sources, the feature invariant to illumination will be also important. To<br />
investigate the effect of illumination, let us consider two iris images from the same<br />
iris image with different illuminations shown in Figs. 11(a) and 11(b). As can be<br />
seen from their EMD decomposition results in Figs. 11(c) and 11(d), their difference<br />
is mainly on the residues. The residue of Fig. 11(c) is about 190, while the residue<br />
of Fig. 11(d) is about 130. As we know the residue of EMD is removed before<br />
computing the Hilbert marginal spectrum, so the illumination variation will not<br />
influence the main frequency center and its energy. If we remove the residues of<br />
all lines from the original images, the residual images of Figs. 11(a) and 11(b)<br />
are shown in Figs. 11(e) and 11(f), respectively. It can be seen that it not only<br />
compensates for the nonuniform illumination but also improves the contrast of the
x<br />
c1<br />
c2<br />
c3<br />
c4<br />
c5<br />
Iris Recognition Based on Hilbert–Huang Transform 635<br />
(a) (b)<br />
(c) (d)<br />
200<br />
150<br />
100<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
50<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−50<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
10<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−10<br />
0<br />
5<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−5<br />
0<br />
180<br />
160<br />
50 100 150 200 250 300 350 400 450 500<br />
140<br />
0 50 100 150 200 250 300 350 400 450 500<br />
c6<br />
r<br />
200<br />
150<br />
100<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
50<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−50<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
10<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−10<br />
0<br />
5<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−5<br />
0<br />
180<br />
160<br />
50 100 150 200 250 300 350 400 450 500<br />
140<br />
0 50 100 150 200 250 300 350 400 450 500<br />
(e) (f)<br />
Fig. 10. (a) Original image; (b) the image after rotating 90 ◦ clockwise; (c) the normalized image<br />
of the original image; (d) the normalized image of the rotated image; (e) the EMD decomposition<br />
result for the 10th line along horizon of the original normalized image; (f) the EMD decomposition<br />
result for the 10th line along horizon of the rotated normalized image.<br />
image. Therefore other than most of other iris recognition methods, our method<br />
need not do the enhancement processing in the iris image preprocessing.<br />
3.3.2. Robustness to noise<br />
Just as stated in Refs. 3, 11 and 26 the high frequency noise is mainly contained in<br />
the first IMF. Moreover, it is found that the first IMF is not a significant component<br />
to characterize the iris structure. To be robust to high frequency noise, the first IMF<br />
is removed when the Hilbert marginal spectrum is calculated in our method. Though<br />
it leads a slight change for the main frequency, it can relieve the high frequency<br />
noises.<br />
x<br />
c1<br />
c2<br />
c3<br />
c4<br />
c5<br />
c6<br />
r
636 Z.Yang,Z.Yang&L.Yang<br />
x<br />
c1<br />
c2<br />
c3<br />
c4<br />
c5<br />
c6<br />
r<br />
250<br />
200<br />
150<br />
(a) (b)<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
10<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−10<br />
0<br />
10<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−10<br />
0<br />
200<br />
180<br />
50 100 150 200 250 300 350 400 450 500<br />
160<br />
0 50 100 150 200 250 300 350 400 450 500<br />
x<br />
c1<br />
c2<br />
c3<br />
c4<br />
c5<br />
c6<br />
r<br />
200<br />
150<br />
100<br />
50<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
20<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−20<br />
0<br />
10<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−10<br />
0<br />
10<br />
0<br />
50 100 150 200 250 300 350 400 450 500<br />
−10<br />
0<br />
150<br />
50 100 150 200 250 300 350 400 450 500<br />
100<br />
0 50 100 150 200 250 300 350 400 450 500<br />
(c) (d)<br />
(e) (f)<br />
Fig. 11. (a) Iris image with bright illumination; (b) the same iris image with dark illumination;<br />
(c) the EMD decomposition result of the 10th line of (a); (d) the EMD decomposition result of<br />
the 10th line of (b); (e) the result by removing all the residues of lines from (a); (f) the result by<br />
removing all the residues of lines from (b).<br />
To show the robustness of our method to noise, we added 20 dB Gauss white<br />
noise to the iris image as shown in Fig. 12(a) and calculate the main frequency<br />
centers and the energies of 18 orientations in I1 of the original normalized image<br />
and the noise normalized image, as shown in Figs. 12(b) and 12(c), respectively.<br />
It can be seen that most features of the noisy iris image just have small changes<br />
compared with those of the original iris. Therefore, the proposed feature is robust<br />
to high frequency noise.<br />
3.4. Iris matching<br />
After feature extraction, an iris image is represented as a feature vector of length<br />
108. To improve computational efficiency and classification accuracy, Linear Discriminant<br />
Analysis (LDA) is first used to reduce the dimensionality of the feature<br />
vector and then the Euclidean similarity measure is adopted for classification. LDA<br />
is a linear statistic classification method, which intends to find a linear transform T<br />
as such that, after its application, the scatter of sample vectors is minimized within<br />
each class, and the scatter of those mean vectors around the total mean vector can<br />
be maximized simultaneously. Further details of LDA may be found in Ref. 8.
Main Frequency Center<br />
0.06<br />
0.05<br />
0.04<br />
0.03<br />
0.02<br />
0 20 40 60 80<br />
Orientation<br />
100 120 140 160<br />
Iris Recognition Based on Hilbert–Huang Transform 637<br />
Original features<br />
Noise features<br />
(a)<br />
Energy<br />
500<br />
400<br />
300<br />
200<br />
Original features<br />
Noise features<br />
0 20 40 60 80<br />
Orientation<br />
100 120 140 160<br />
(b) (c)<br />
Fig. 12. (a) The original iris image; (b) the main frequency centers of 18 orientations in I1 of the<br />
original normalized image (“o”) and those of the noisy normalized image (“•”); (c) the energies<br />
of 18 orientations in I1 of the original normalized image (“o”) and those of the noisy normalized<br />
image (“•”).<br />
4. Experimental Results<br />
To evaluate the performance of the proposed method, we applied it to the widely<br />
used database named CASIA iris database. 1 The database includes 2255 iris images<br />
from 306 different eyes (hence, 306 different classes). The captured iris images are<br />
8-bit gray images with a resolution of 320 × 280.<br />
4.1. Performance of the proposed method<br />
For each iris class, we choose three samples taken at the first session for training and<br />
all samples captured at the second and third sessions serve as test samples. Therefore,<br />
there are 918 images for training and 1337 images for testing. Figure 13(a)<br />
describes variations of the correct recognition rate (CRR) with changes of dimensionality<br />
of the reduced feature vector using the LDA. From this figure, we can see<br />
that with increasing dimensionality of the reduced feature vector, the recognition<br />
rate also increases rapidly. However, when the dimensionality of the reduced feature
638 Z.Yang,Z.Yang&L.Yang<br />
Correct recognition rate<br />
100<br />
99.5<br />
99<br />
98.5<br />
98<br />
97<br />
96<br />
95<br />
94<br />
93<br />
92<br />
91<br />
10 20 30 40 50 60 70 80 90<br />
Dimensionality of the feature vector<br />
Density (£¥)<br />
40<br />
35<br />
30<br />
25<br />
20<br />
15<br />
10<br />
5<br />
intra-class distribution<br />
inter-class distribution<br />
0<br />
0 5 10 15 20 25 30 35 40<br />
Normalized matching distance<br />
(a) (b)<br />
Fig. 13. (a) Recognition results using features of different dimensionality using LDA. (b) Distributions<br />
of intra- and inter-class distances.<br />
vector is up to 70 or higher, the recognition rate starts to level off at an encouraging<br />
CRR of about 99.4%. In the experiments, we utilize the Euclidian similarity<br />
measure for matching. The results indicate that the proposed method will be highly<br />
feasible in practical applications.<br />
To evaluate the performance of the proposed method in verification mode,<br />
each tested iris image is compared with all the trained iris images on the CASIA<br />
database. Therefore, the total number of comparisons is 1337 × 918 = 1,227,366,<br />
where the total number of intra-class comparisons is 1337 × 3 = 4011, and that<br />
of inter-class comparisons is 1337 × 915 = 1,223,355. Figure 13(b) shows distributions<br />
of intra- and inter-class matching distances on the CASIA iris database.<br />
As shown in Fig. 13(b), we can find that the distance between the intra- and the<br />
inter-class distributions is large, and the portion that overlaps between the intraand<br />
the inter-class is very small. This proves that the proposed features are highly<br />
discriminating.<br />
In verification mode, the receiver operating characteristic (ROC) curve and<br />
equal error rate (EER) are used to evaluate the performance of the proposed<br />
method. 17 The ROC curve is a false accept rate (FAR) versus false reject rate<br />
(FRR) curve, which measures the accuracy of matching process and shows the<br />
overall performance of an algorithm. Points on this curve denote all possible system<br />
operating states in different tradeoffs. The EER is the point where the FAR<br />
and the FRR are equal in value. The smaller the EER is, the better the algorithm<br />
is. Figure 14(a) shows the ROC curve of the proposed method on the CASIA iris<br />
databases. It can be seen that the performance of our algorithm is very high and<br />
the EER is only 0.27%. If we choose the threshold as 17.2, the CRR is up to 99.7%<br />
when the FRR is less than 0.15%.
False Reject Rate(%)<br />
0<br />
2.5<br />
2<br />
1.5<br />
1<br />
0.5<br />
0.1 0.2 0.3<br />
False Accept Rate(%)<br />
0.4 0.5<br />
Iris Recognition Based on Hilbert–Huang Transform 639<br />
Correct recognition rate<br />
100<br />
90<br />
80<br />
70<br />
60<br />
50<br />
40<br />
30<br />
20<br />
10<br />
0<br />
40dB 30dB 20dB<br />
Signal−to−noise ratio<br />
10dB<br />
(a) (b)<br />
Fig. 14. (a) The ROC curve of our method on CASIA iris database. (b) Recognition results of<br />
adding different SNR to the tested iris images.<br />
As discussed in the above section, our method is robust to high frequency noise.<br />
To evaluate the robustness of our method to noise, we added Gauss white noise with<br />
different signal-to-noise ratio (SNR) to the test iris images. The result is shown in<br />
Fig. 14(b). The result is encouraging, since the correct recognition rates are 97.2%,<br />
95.6%, 92.2% and 87.5%, respectively, when SNR = 40, 30, 20 and 10 dB.<br />
4.2. Comparison with existing methods<br />
Among existing methods for iris recognition, those proposed by Daugman, 7 Wildes<br />
et al., 4 and Tan et al., 14,15 respectively, are the best known. Furthermore, they<br />
characterize the iris from different viewpoints. To further prove the effectiveness<br />
of our method, we make comparison with the above four methods on the CASIA<br />
iris database. The experimental results are shown in the Table 1. It can be seen<br />
that the performance of our method is encouraging and comparable to the best<br />
iris recognition algorithm while the dimension of our feature vector is very low.<br />
Furthermore, as discussed above our method is rotation-invariant and illuminationinvariant.<br />
Besides it is robust to high frequency noise.<br />
Table 1. The feature dimension and correct recognition rate comparisons<br />
of several well-known methods.<br />
Methods Feature dimension Correct recognition rate (%)<br />
Boles and Boashash 4 1024 92.4<br />
Daugman 7 2048 100<br />
Tan and co-workers 15 660 99.6<br />
Tan and co-workers 14 1536 99.43<br />
Our method 108 99.4
640 Z.Yang,Z.Yang&L.Yang<br />
5. Conclusions<br />
Recently, iris recognition has received increasing attention for human identification<br />
due to its high reliability. HHT is an analysis method for nonlinear and<br />
nonstationary data. In this paper, we have presented an efficient iris recognition<br />
method based on HHT by extracting the main frequency center information. This<br />
new method benefits a lot: firstly, its dimension of feature vector is very low compared<br />
with the other famous methods; secondly, other than most of other iris<br />
recognition methods, our method need not do the enhancement processing in the<br />
iris image preprocessing and is illumination-invariant; thirdly, unlike most existing<br />
methods to achieve approximate rotation invariance by defining several templates<br />
denoting other angles, our method is really rotation-invariant; fourthly, it is robust<br />
to high frequency noise; Moreover, the experimental results on the CASIA iris<br />
database show that the correct recognition rate of the proposed method is encouraging<br />
and comparable to the best iris recognition algorithm. In addition, the proposed<br />
method has demonstrated that HHT is a powerful tool for feature extraction<br />
and will be useful for many other pattern recognitions.<br />
Acknowledgments<br />
Portions of the research in this paper use the CASIA iris image database collected<br />
by Institute of Automation, Chinese Academy of Sciences.<br />
This work is supported by NSFC (Nos. 10631080, 60873088, 60475042), and<br />
NSFGD (No. 9451027501002552).<br />
References<br />
1. CASIA iris image database, http://www.sinobiometrics.com.<br />
2. F. H. Adler, Physiology of the Eye (Mosby, 1965).<br />
3. N. Bi, Q. Sun, D. Huang, Z. Yang and J. Huang, Robust image watermarking based on<br />
multiband wavelets and empirical mode decomposition, IEEE Trans. Image Process.<br />
16(8) (2007) 1956–1966.<br />
4. W. Boles and B. Boashash, A human identification technique using images of the iris<br />
and wavelet transform, IEEE Trans. Signal Process. 46(4) (1998) 1185–1188.<br />
5. C.-P. Chang, J.-C. Lee, Y. Su, P. S. Huang and T.-M. Tu, Using empirical<br />
mode decomposition for iris recognition, Comput. Standards Interfaces 31(4) (2009)<br />
729–739.<br />
6. J. Daugman, High confidence visual recognition of persons by a test of statistical<br />
independence, IEEE Trans. Pattern Anal. Mach. Intell. 15(11) (1993) 1148–1161.<br />
7. J. Daugman, Statistical richness of visual phase information: Update on recognizing<br />
persons by iris patterns, Int. J. Comput. Vision 45(1) (2001) 25–38.<br />
8. R. A. Fisher, The use of multiple measures in taxonomic problems, Ann. Eugenics 7<br />
(1936) 179–188.<br />
9. P. Flandrin, G. Rilling and P. Goncalves, Empirical mode decomposition as a filter<br />
bank, IEEE Signal Process. Lett. 11(2) (2004) 112–114.<br />
10. N. E. Huang, Z. Shen and S. R. Long, A new view of nonlinear water waves: The<br />
Hilbert spectrum, Ann. Rev. Fluid Mechanics 31 (1999) 417–457.
Iris Recognition Based on Hilbert–Huang Transform 641<br />
11. N. E. Huang, Z. Shen, S. R. Long, M. C. Wu, H. H. Shih, Q. Zheng, N.-C. Yen, C. C.<br />
Tung and H. H. Liu, The empirical mode decomposition and the Hilbert spectrum<br />
for nonlinear and non-stationary time series analysis, Proc. Roy. Soc. London A 454<br />
(1998) 903–995.<br />
12. W. Huang, Z. Shen, N. E. Huang and Y. C. Fung, Engineering analysis of biological<br />
variables: An example of blood pressure over 1 day, Proc. Natl. Acad. Sci. USA 95<br />
(1998) 4816–4821.<br />
13. A. Jain, R. Bolle and S. Pankanti, Biometrics: Personal Identification in Networked<br />
Society (Kluwer, Norwell, MA, 1999).<br />
14. L. Ma, T. Tan, Y. Wang and D. Zhang, Personal identification based on iris texture<br />
analysis, IEEE Trans. Patt. Anal. Mach. Intell. 25(12) (2003) 1519–1533.<br />
15. L. Ma, T. Tan, Y. Wang and D. Zhang, Efficient iris recognition by characterizing<br />
key local variations, IEEE Trans. Image Process. 13(6) (2004) 739–750.<br />
16. L. Ma, T. Tan, D. Zhang and Y. Wang, Local intensity variation analysis for iris<br />
recognition, Patt. Recogn. 37(6) (2005) 1287–1298.<br />
17. A. Mansfield and J. Wayman, Best Practices in Testing and Reporting Performance<br />
of Biometric Devices (National Physical Laboratory of UK, 2002).<br />
18. S. Meignen and V. Perrier, A new formulation for empirical mode decomposition based<br />
on constrained optimization, IEEE Signal Process. Lett. 14(12) (2007) 932–935.<br />
19. J. C. Nunes, S. Guyot and E. Delechelle, Texture analysis based on local analysis of<br />
the bidimensional empirical mode decomposition, Machine Vis. Appl. 16(3) (2005)<br />
177–188.<br />
20. S. C. Phillips, R. J. Gledhill, J. W. Essex and C. M. Edge, Application of the Hilbert–<br />
Huang transform to the analysis of molecular dynamic simulations, J. Phys. Chem.<br />
A 107 (2003) 4869–4876.<br />
21. G. Rilling and P. Flandrin, One or two frequencies? The empirical mode decomposition<br />
answers, IEEE Trans. Signal Process. 56(1) (2007) 85–95.<br />
22. C. Sanchez-Avila and R. Sanchez-Reillo, Iris-based biometric recognition using dyadic<br />
wavelet transform, IEEE Aerosp. Electron. Syst. Mag. 17 (2002) 3–6.<br />
23. C. Tisse, L. Martin, L. Torres and M. Robert, Person identification technique using<br />
human iris recognition, in Proc. Vision Interface (2002), pp. 294–299.<br />
24. R.Wildes,J.Asmuth,G.Green,S.Hsu,R.Kolczynski,J.MateyandS.McBride,A<br />
machine-vision system for iris recognition, Mach. Vision Appl. 9 (1996) 1–8.<br />
25. Z. Yang, D. Huang and L. Yang, A novel pitch period detection algorithm based on<br />
Hilbert–Huang transform, Lecture Notes Comput. Sci. 3338 (2004) 586–593.<br />
26. Z. Yang, D. Qi and L. Yang, Signal period analysis based on Hilbert–Huang transform<br />
and its application to texture analysis, in Proceedings of the Third International Conference<br />
on Image and Graphics, Hong Kong, China (18–20 Dec. 2004), pp. 430–433.<br />
27. Z. Yang and L. Yang, A novel texture classification method using multidirections<br />
main frequency center, in Proceedings of the 2007 International Conference<br />
on Wavelet Analysis and Pattern Recognition, Beijing, China (2–4 Nov. 2007),<br />
pp. 1372–1376.<br />
28. Z. Yang, L. Yang, D. Qi and C. Y. Suen, An EMD-based recognition method for<br />
Chinese fonts and styles, Patt. Recogn. Lett. 27 (2006) 1692–1701.<br />
29. D. Zhang, Automated Biometrics: Technologies and Systems (Kluwer, Norwell, MA,<br />
2000).