A survey paper on various biometric security system methods
Face Recognition report
1. 2013
BIOMETRICS
Facial Recognition System
Developing an algorithm for facial recognition
system using Image processing in Matlab
T. Lavanya,EEE,NITK
Under the guidance of Professor DR. K.B.Raja, Department of Electronics and
Communication, UVCE, Bangalore.
2. 1
1.INTRODUCTION
Biometrics:
Biometrics or biometric authentication refers to the identification of
humans by their characteristics or traits (distinguishing features). It is
used as a form of identification, and access control, also to identify
individuals in groups that are under surveillance. Biometric identifiers
are the distinctive, measurable characteristics used to label and describe
individuals. They are often categorized as physiological versus
behavioral characteristics.
1. Physiological characteristics are related to the shape of the body.
Examples include fingerprint, face recognition, DNA, Palm print,
hand geometry, iris recognition, retina and odour /scent.
2. Behavioral characteristics are related to the behavior of a person,
including typing rhythm, gait, and voice.
Behaviometrics has been coined by researchers to describe the latter
class of biometrics.
What is the need for Biometrics?
Access control through token based identification systems, such as a
driver's license or passport, and knowledge based identification systems,
such as a password or personal identification number have become
traditional ways and turned to be inefficient. Since biometric identifiers
are unique to individuals, they are more reliable in verifying identity
than token and knowledge based methods leads to opt biometrics
3. 2
How does Biometrics function?
Different aspects of human physiology, chemistry or behavior can be
used for biometric authentication. The selection of a particular biometric
for use in a specific application involves a weighting of several factors.
Jain et al. (1999) has identified seven such factors to be used when
assessing the suitability of any trait for use in biometric authentication.
1. Universality means that every person using a system should
possess the trait.
2. Uniqueness means the trait should be sufficiently different for
individuals in the relevant population such that they can be
distinguished from one another.
3. Permanence relates to the manner in which a trait varies over
time. More specifically, a trait with 'good' permanence will be
reasonably invariant over time with respect to the specific
matching algorithm.
4. Measurability (collectability) relates to the ease of acquisition or
measurement of the trait. In addition, acquired data should be in a
form that permits subsequent processing and extraction of the
relevant feature sets.
5. Performance relates to the accuracy, speed, and robustness of
technology used.
6. Acceptability relates to how well individuals in the relevant
population accept the technology such that they are willing to have
their biometric trait captured and assessed.
7. Circumvention relates to the ease with which a trait might be
imitated using an artifact or substitute.
No single biometric will meet all the requirements of every
possible application.
4. 3
Basic block diagram of a biometric system in 2 modes
In verification mode the system performs a one-to-one comparison of a
captured biometric with a specific template stored in a biometric
database in order to verify the individual is the person they claim to be.
Three steps involved in person verification.
In the first step, reference models for all the users are generated and
stored in the model database.
In the second step, some samples are matched with reference models to
generate the genuine and impostor scores and calculate the threshold.
Third step is the testing step. This process may use a smart card,
username or ID number (e.g. PIN) to indicate which template should be
used for comparison. 'Positive recognition' is a common use of
verification mode, "where the aim is to prevent multiple people from
using same identity".
5. 4
The first time an individual uses a biometric system is called enrollment.
During the enrollment, biometric information from an individual is
captured and stored. In subsequent uses, biometric information is
detected and compared with the information stored at the time of
enrollment. Note that it is crucial that storage and retrieval of such
systems themselves be secure if the biometric system is to be robust.
The first block (sensor) is the interface between the real world and the
system; it has to acquire all the necessary data.
The second block performs all the necessary pre-processing: it has to
remove artifacts from the sensor, to enhance the input (e.g. removing
background noise), to use some kind of normalization, etc.
In the third block necessary features are extracted. This step is an
important step as the correct features need to be extracted in the optimal
way. A vector of numbers or an image with particular properties is used
to create a template. A template is a synthesis of the relevant
characteristics extracted from the source.
Elements of the biometric measurement that are not used in the
comparison algorithm are discarded in the template to reduce the file
size and to protect the identity of the enrollee.
If enrollment is being performed, the template is simply stored
somewhere (within a database). If a matching phase is being performed,
the obtained template is passed to a matcher that compares it with other
existing templates, estimating the distance between them using any
algorithm (e.g. Hamming distance). The matching program will analyze
the template with the input. This will then be output for any specified
use or purpose (e.g. entrance in a restricted area)
6. 5
Performance:
The following are used as performance metrics for biometric systems:
False Accept Rate Or False Match Rate (FAR or FMR): The
probability that the system incorrectly matches the input pattern to a
non-matching template in the database. It measures the percent of
invalid inputs which are incorrectly accepted. In case of similarity scale,
if the person is imposter in real, but the matching score is higher than the
threshold then he is treated as genuine and that increases the FAR and
hence performance also depends upon the selection of threshold value.
False Reject Rate Or False Non-Match Rate (FRR Or FNMR): The
probability that the system fails to detect a match between the input
pattern and a matching template in the database. It measures the percent
of valid inputs which are incorrectly rejected.
Receiver Operating Characteristic Or Relative Operating
Characteristic (ROC): The ROC plot is a visual characterization of the
trade-off between the FAR and the FRR. In general, the matching
algorithm performs a decision based on a threshold which determines
how close to a template the input needs to be for it to be considered a
match. If the threshold is reduced, there will be fewer false non-matches
but more false accepts. Correspondingly, a higher threshold will reduce
the FAR but increase the FRR. A common variation is the Detection
error trade-off (DET), which is obtained using normal deviate scales on
both axes. This more linear graph illuminates the differences for higher
performances (rarer errors).
Equal Error Rate Or Crossover Error Rate (EER Or CER): The rate
at which both accept and reject errors are equal. The value of the EER
can be easily obtained from the ROC curve. The EER is a quick way to
7. 6
compare the accuracy of devices with different ROC curves. In general,
the device with the lowest EER is most accurate.
Failure To Enroll Rate (FTE Or FER): The rate at which attempts to
create a template from an input is unsuccessful. This is most commonly
caused by low quality inputs.
Failure To Capture Rate (FTC): Within automatic systems, the
probability that the system fails to detect a biometric input when
presented correctly.
Template Capacity: The maximum number of sets of data which can
be stored in the system.
Selection of biometrics in any practical application depending upon the
characteristic measurements and user requirements We should consider
Performance, Acceptability, Circumvention, Robustness, Population
coverage, Size, Identity theft deterrence in selecting a particular
biometric. Selection of biometric based on user requirement considers
Sensor availability, Device availability, Computational time and
reliability, Cost, Sensor area and power consumption.
As a proverb says, “face is the index of mind”, face is the first and
foremost feature in a human being which distinguishes one from
another. A person is identified or remembered by his face which is
unique for every person (except in some cases). Using biometrics of
facial recognition is preferred since many systems which require
access control has the identity as the image of face of the person rather
than signature, palm or finger print and if surveillance cameras are
being used it is easy to capture ones face without the person knowing
it, which is not possible with signature, palm or finger print.
8. 7
2. FACIAL RECOGNITION SYSTEM
A facial recognition system is a computer application for
automatically identifying or verifying a person from a digital image
or a video frame from a video source. One of the ways to do this is by
comparing selected facial features from the image and a facial
database. It is typically used in security systems.
Facial recognition algorithms identify facial features by extracting
landmarks, or features, from an image of the subject's face. For
example, an algorithm may analyze the relative position, size, and/or
shape of the eyes, nose, cheekbones, and jaw. These features are then
used to search for other images with matching features. Other
algorithms normalize a gallery of face images and then compress the
face data, only saving the data in the image that is useful for face
detection. A probe image is then compared with the face data. One of
the earliest successful systems is based on template matching
techniques applied to a set of salient facial features, providing a sort
of compressed face representation. Recognition algorithms can be
divided into two main approaches, geometric, which look at
distinguishing features, or photometric, which is a statistical approach
that distills an image into values and compares the values with
templates to eliminate variances.
In the present world face recognition finds it use in every walk of life.
Starting with smart phones, digital cameras, tabs, laptops and
televisions, all are coming with this feature of face recognition
enabling their use in more personalized way. Almost every place
either a public place like banks, airports, government offices,
universities or private (business) places like shopping malls are being
watched by surveillance cameras which can be used for tracking a
suspicious person or in general anyone using face recognition.
9. 8
Google's Picasa digital image organizer has a built in face recognition
system.
Sony's Picture Motion Browser (PMB) analyses photo, associates
photos with identical faces so that they can be tagged accordingly,
and differentiates between photos with one person, many persons and
nobody.
Windows Live Photo Gallery includes face recognition
Apple's iPhoto image organizer has a feature named Faces which
allows users to associate names to faces in a series of photographs.
The software then proposes other matching faces it locates in the
photo library and gradually refines its recognition, according to the
user's acceptance, denial or renaming of the proposed face.
Even social networking sites like facebook use tagging photos of a
person based on this facial recognition
This technology could be used as a security measure at ATM’s; where
the ATM would capture an image of your face, and compare it to your
photo in the bank database to confirm your identity.
This same concept is being used in computers by using a webcam to
capture a digital image of your face which replaces the necessity of
password as a means to log-in.
Another use could be a portable device to assist people with
prosopagnosia in recognizing their acquaintances.
COMPARITIVE STUDY: Though facial recognition is used widely,
among the different biometric techniques, facial recognition may not be
the most reliable and efficient. However, one key advantage is that it
does not require aid (or consent) from the test subject. Properly designed
systems installed in airports, multiplexes, and other public places can
identify individuals among the crowd. Other biometrics like fingerprints,
iris scans, and speech recognition cannot perform this kind of mass
identification. However, questions have been raised on the effectiveness
of facial recognition software in cases of railway and airport security.
10. 9
Two algorithms are being used in this project, first is using preprocessed
image, i.e. comparison of images in spatial domain by considering the
pixel values.
Algorithm 1: Spatial domain
In general the face recognition is carried out in four phases
a) Image Acquisition
b) Preprocessing
c) Feature Extraction and
d) Matching
Database Face images Text face images
Resize (128*128)
Co-efficient vector
Euclidian Distance classifier
Face image match/non match
a) Image Acquisition: Images are acquired from open source
databases like ORL, JAFFE, Indian male, Indian female and more.
Images in each database consist of various images of each person
taken at different modes such as variation in pose, illumination,
facial expressions and with different accessories
11. 10
i) At variety of pose angles to permit testing of pose invariance
ii) With a wide variety of illumination angles to permit testing of
illumination invariance and
iii) Under a variety of commonly encountered illumination color
temperatures to permit testing of illumination color invariance.
b) Preprocessing:
Preprocessing is done to extract the features of the image with minimal
difficulties such as resizing the image. Resizing makes all the images in
the database into a particular size which is common to all the images.
c) Feature Extraction:
From the preprocessed images features are extracted, in this project, the
pixel values of each image are stored in a vector and all the images to be
considered in the database are stored in the form of vector of size equal
to the total number of images.
If a database is considered with n persons and m images of each person,
a database should be created which contains x (say x<n) persons and y
(say y<m) images of each person and the remaining persons and images
are considered for test database.
The first three steps are common for both database and test database.
d) Matching:
The features of database and test database are compared by using a
classifier which in this case is Euclidean distance. Each test image is
compared with every image present in the database. The Euclidean
distance is zero if same image is considered. If test image belongs to a
person whose images are already there in database, then the resulting
Euclidean distance should be minimum or else it should be maximum.
Implementation using Matlab:
Based on the above algorithm, a matlab code has been developed to
calculate, plot the performance parameters namely FAR,FRR,TSR and
obtain EER from the results. JAFFE Database is being considered.
Results for one test case are shown below:
No of persons in database=7; No of persons in test database=10;
No of images of each person in database=12; Test image=17;
Image size = 128*128;
14. 13
The EER is 0.142857 corresponding to the optimum threshold of
0.449611
The TSR is 0.857143 corresponding to the optimum threshold of
0.449611
Database: JAFFE
The following table gives the details of results when the same algorithm
and Matlab code applied on various test images by changing number of
persons in database, number of images of each person and test image.
No. of
persons
in
database
No. of
persons
out of
database
Number
of
images
of each
person
Number
of the
test
image
EER TSR Optimum
threshold
preprocessed
5 5 15 17 0.100000 0.900000 0.432737
5 5 15 18 0.000000 1.000000 0.411111
5 5 15 19 0.200000 0.800000 0.455770
5 5 15 20 0.200000 0.800000 0.458616
7 3 15 17 0.000000 1.000000 0.444960
7 3 15 18 0.000000 1.000000 0.459256
7 3 15 19 0.142857 0.857143 0.462261
7 3 15 20 0.142857 0.857143 0.482867
6 4 15 17 0.000000 1.000000 0.452415
6 4 15 18 0.000000 1.000000 0.441222
6 4 15 19 0.181818 0.818182 0.455621
6 4 15 20 0.166667 0.833333 0.464738
7 3 12 13 0.000000 1.000000 0.445503
7 3 12 15 0.000000 1.000000 0.477148
7 3 12 17 0.142857 0.857143 0.452712
It has been observed the algorithm gives EER values in the range of 0.20
to 0.00 and TSR values in range of 0.80 to 1.00 for this database.
15. 14
Algorithm 2: Using DWT on preprocessed image (Discrete wavelet
transform)
Discrete Wavelet Transform:
The transform of a signal is just another form of representing the signal.
It does not change the information content present in the signal. The
Wavelet Transform provides a time-frequency representation of the
signal. It was developed to overcome the short coming of the Short Time
Fourier Transform (STFT), which can also be used to analyze non
stationary signals. While STFT gives a constant resolution at all
frequencies, the Wavelet Transform uses multi-resolution technique by
which different frequencies are analyzed with different resolutions.
A wave is an oscillating function of time or space and is periodic. In
contrast, wavelets are localized waves. The figure shown below gives
the demonstration of a wave and a wavelet. They have their energy
concentrated in time or space and are suited to analysis of transient
signals. While Fourier Transform and STFT use waves to analyze
signals, the Wavelet Transform uses wavelets of finite energy.
Demonstration of Wave
16. 15
Demonstration of Wavelet
The wavelet analysis is done similar to the STFT analysis. The signal to
be analyzed is multiplied with a wavelet function just as it is multiplied
with a window function in STFT, and then the transform is computed for
each segment generated. However, unlike STFT, in Wavelet Transform,
the width of the wavelet function changes with each spectral component.
The Wavelet Transform, at high frequencies, gives good time resolution
and poor frequency resolution, while at low frequencies the Wavelet
Transform gives good frequency resolution and poor time resolution.
The Wavelet Series is just a sampled version of CWT and its
computation may consume significant amount of time and resources,
depending on the resolution required. The Discrete Wavelet Transform
(DWT), which is based on sub-band coding is found to yield a fast
computation of Wavelet Transform. It is easy to implement and reduces
the computation time and resources required. In CWT, the signals are
analyzed using a set of basis functions which relate to each other by
simple scaling and translation. In the case of DWT, a time-scale
representation of the digital signal is obtained using digital filtering
techniques. The signal to be analyzed is passed through filters with
different cutoff frequencies at different scales
17. 16
One Dimensional Discrete Wavelet Transform
Two-Channel Perfect Reconstruction Filter Bank:
The analysis filter bank decomposes the input signal x (n) into two sub
band signals, L (n) and H (n). The signal L (n) represents the low
frequency (coarse) part of x (n), while the signal H (n) represents the
high frequency (or detail) part of x (n). The analysis filter bank first
filters x (n) using a low pass and a high pass filter. We denote the low
pass filter by af1 (analysis filter 1) and the high pass filter by af2
(analysis filter 2). As shown in the figure 3.2, the output of each filter is
then down-sampled by 2 to obtain the two sub band signals, L (n) and
H (n).
Reconstruction filter band
The synthesis filter bank combines the two sub band signals L (n) and
H (n) to obtain a single signal y(n). The synthesis filter bank first up-
samples each of the two sub band signals. The signals are then filtered
using a low pass and a high pass filter. We denote the low pass filter by
sf1 (synthesis filter 1) and the high pass filter by sf2 (synthesis filter 2).
The signals are then added together to obtain the signal y (n). If the four
filters are designed so as to guarantee that the output signal y (n) equals
the input signal x (n), then the filters are said to satisfy the perfect
reconstruction condition.
18. 17
Assume the input signal x (n) is of length N. For convenience, we will
like the sub-band signals L (n) and H (n) to each be of length N/2.
However, these sub-band signals will exceed this length by L1/2, where
L1 is the length of the analysis filters. To avoid this excessive length, the
last L1/2 samples of each sub-band signal is added to the first L1/2
samples. This procedure (periodic extension) can create undesirable
artifacts at the beginning and end of the sub-band signals, however, it is
the most convenient solution. When the analysis and synthesis filters are
exactly symmetric, a different procedure (symmetric extension) can be
used, that avoids the artifacts associated with periodic extension. A
second detail also arises in the implementation of the perfect
reconstruction filter bank. If all four filters are causal, then the output
signal y(n) will be a translated (or circularly shifted) version of x(n). To
avoid this, we perform a circular shift operation in both the analysis and
synthesis filter banks.
Discrete Wavelet Transform (Iterated Filter Banks)
The Discrete Wavelet Transform (DWT) gives a multiscale
representation of a signal x(n). The DWT is implemented by iterating
the 2-channel analysis filter bank described above. Specially, the DWT
of a signal is obtained by recursively applying the lowpass/highpass
frequency decomposition to the lowpass output as illustrated in the
diagram 3.3. The diagram illustrates a 3-scale DWT. The DWT of the
signal x is the collection of subband signals. The inverse DWT is
obtained by iteratively applying the synthesis filter bank.
19. 18
2-Dimensional Discrete Wavelet Transform
2-D Filter Banks
To use the wavelet transform for image processing we must implement a
2D version of the analysis and synthesis filter banks. In the 2D case, the
1D analysis filter bank is first applied to the columns of the image and
then applied to the rows. If the image has M rows and N columns, then
after applying the 1D analysis filter bank to each column we have two
sub-band images, each having M/2 rows and N columns; after applying
the 1D analysis filter bank to each row of both of the two sub-band
images, we have four sub-band images, each having M/2 rows and N/2
columns. This is illustrated in the diagram below. The 2D synthesis filter
bank combines the four sub-band images to obtain the original image of
size M by N.
One stage in multi-resolution wavelet decomposition of an image
2D Discrete Wavelets
Like in the 1D case, the 2D discrete wavelet transform of a signal x is
implemented by iterating the 2D analysis filter bank on the low pass
20. 19
sub-band image. In this case, at each scale there are three sub-bands
instead of one.
There are three wavelets associated with the 2D wavelet transform. The
following figure illustrates three wavelets as gray scale images.
Note that the first two wavelets are oriented in the vertical and
horizontal directions; however, the third wavelet does not have a
dominant orientation. The third wavelet mixes two diagonal orientations,
which gives rise to the checkerboard artifact.
Advantages of Discrete wavelet transform
1. It gives information about both time and frequency of the signal.
2. Transform of a non stationary signal is efficiently obtained.
3. Reduces the size without losing much of resolution
4. Reduces redundancy.
5. Reduces computational time
Disadvantages of DWT:
1. Lack of shift invariance.
2. Lack of directional selectivity for higher dimensionality.
3. Unsatisfactory reconstruction.
4. It has more redundancy compare to DTCWT
21. 20
The first two phases a) Image acquisition and b)Preprocessing are same
as in algorithm1but extracting features is done after applying 2-D DWT
on preprocessed image and the images are stored in database and test
database and matching has to be done.
Based on this, performance parameters have to calculated and results are
plotted.
2-D DWT gives four compressed band namely LL, LH, HL and HH.
Among these LL band is preferred since it gives approximate and better
results compared to LH, HL and HH bands. Haar wavelet is being used
in DWT.
Database Face images Text face images
Resize (128*128)
Discrete Wavelet Transform
LL Sub band
Co-efficient vector
Euclidian Distance classifier
Face image match/non match
22. 21
Implementation with matlab: 2-D Dwt can be calculated in matlab
using an inbuilt command by choosing required wavelet/ wave filter.
In this test case2-D Dwt is calculated using ‘haar’ wavelet for JAFFE
Database choosing No of persons in database=7;
No of persons in test database=10;
No of images of each person in database=15; Test image=19
FRR versus Threshold
24. 23
The EER is 0.200000 corresponding to the optimum threshold of
0.432602
The TSR is 0.800000 corresponding to the optimum threshold of
0.432602
25. 24
DATABASE: JAFFE
No. of
persons in
database
No. of
persons
out of
database
Number
of images
of each
person
Number
of the test
image
Optimum
threshold
2-D DWT
EER TSR
5 5 15 17 0.408196 0.100000 0.900000
5 5 15 18 0.383855 0.000000 1.000000
5 5 15 19 0.429051 0.200000 0.800000
5 5 15 20 0.432277 0.200000 0.800000
7 3 15 17 0.428339 0.000000 1.000000
7 3 15 18 0.431603 0.000000 1. 000000
7 3 15 19 0.432602 0.200000 0.800000
7 3 15 20 0.462472 0.142857 0.857143
6 4 15 17 0.426365 0.000000 1.000000
6 4 15 18 0.414367 0.000000 1.000000
6 4 15 19 0.432112 0.166667 0.833333
6 4 15 20 0.435446 0.166667 0.833333
7 3 12 20 0.473635 0.142857 0.857143
7 3 12 13 0.429423 0.000000 1.000000
7 3 12 15 0.450294 0.000000 1.000000
7 3 12 16 0.471678 0.000000 1.000000
7 3 12 18 0.431603 0.000000 1.000000
7 3 12 17 0.431552 0.100000 0.900000
7 3 12 19 0.441888 0.142857 0.857143
By varying the number of persons in database, number of images of each
person and test images, the values of EER and TSR are being calculated.
EER is the point of intersection of FAR and FRR at optimal
threshold, the lower the value of EER, the better is the algorithm
Also higher TSR value at EER shows how adaptable and robust
the algorithm is.
Creating a database and test database with images is demonstrated by
an example, consisting of 3 persons each with 8 images in database
and test database with 9th
image of 4 persons among which 3 are in
database and 1 out of database.
27. 26
Test database:
RESULTS ON DIFFERENT DATABASES:
ORL-FACES
No. of
persons
in
database
No. of
persons
out of
database
Number
of
images
of each
person
Number
of the
test
image
Optimum
threshold
EER TSR
20 20 8 10 0.403162 0.050000 0.950000
20 20 8 9 0.440096 0.100000 0.900000
20 20 7 8 0.426077 0.100000 0.900000
20 20 7 9 0.440096 0.100000 0.900000
20 20 7 10 0.397508 0.050000 0.950000
25 15 7 8 0.392461 0.123077 0.876923
25 15 7 9 0.404316 0.133333 0.866667
25 15 7 10 0.387847 0.075000 0.925000
30 10 7 10 0.371744 0.066667 0.933333
30 10 7 9 0.373902 0.183333 0.816667
30 10 7 8 0.388620 0.114286 0.885714
28. 27
Database: Indian females
No. of
persons
in
database
No. of
persons
out of
database
Number
of
images
of each
person
Number
of the
test
image
Optimum
threshold
EER TSR
10 12 8 10 0.336653 0.100000 0.900000
10 12 8 9 0.305086 0.000000 1.000000
10 12 8 11 0.306075 0.100000 0.900000
12 10 8 10 0.338860 0.166667 0.833333
12 10 8 9 0.305086 0.000000 1.000000
12 10 8 11 0.275736 0.090909 0.909091
15 7 7 8 0.283907 0.133333 0.866667
15 7 7 9 0.314187 0.136364 0.863636
15 7 7 10 0.320892 0.133333 0.866667
15 7 7 11 0.349712 0.133333 0.866667
12 10 7 8 0.282989 0.083333 0.916667
12 10 7 9 0.313031 0.083333 0.916667
12 10 7 10 0.338860 0.166667 0.833333
12 10 7 11 0.307584 0.136364 0.863636
ORL_Faces database: It has been observed the same algorithm gives
EER values in the range of 0.183333 to 0.05000 and TSR values in
range of 0.816667 to 0.950000.
Indian Females database: It has been observed the same algorithm
gives EER values in the range of 0.166667 to 0.00000 and TSR values in
range of 0.833333 to 1.00000.
29. 28
Indian Males
No. of
persons
in
database
No. of
persons
out of
database
Number
of images
of each
person
Number
of the test
image
Optimum
threshold
EER TSR
10 10 7 8 0.355908 0.100000 0.900000
10 10 7 9 0.378676 0.300000 0.700000
10 10 6 9 0.382884 0.250000 0.750000
10 10 6 8 0.373269 0.200000 0.800000
10 10 6 7 0.410954 0.300000 0.700000
15 5 7 8 0.285325 0.600000 0.400000
15 5 7 9 0.424989 0.500000 0.500000
12 8 7 8 0.318291 0.083333 0.916667
12 8 7 9 0.326656 0.431818 0.568182
15 5 5 6 0.325532 0.466667 0.533333
15 5 5 7 0.351491 0.457143 0.542857
15 5 5 8 0.286484 0.600000 0.400000
15 5 5 9 0.305127 0.500000 0.500000
The same algorithm is being applied on different databases such as
ORL_Faces, Indian Male and Indian female. Better results are observed
in case of ORL_Faces and not so good results on Indian Males.
Improvement of the algorithm can be done by applying more levels of
DWT or using advanced techniques like DTCWT i.e. Dual Tree
Continuous Wavelet Transform and Complex DTCWT.
30. 29
The following table gives the results of a test case when DWT is applied
using different filters are used such as haar, Daubechies, Coiflets,
Symlets and Discrete Meyer. The values of some of the filters are
similar and that is due to the similarity in the wave function of the
wavelet. The feature size of the DWT may differ for different wavelet
like symlets give feature size of 129 whereas all Daubechies give
feature size as half of the preprocessed image size
Wavelet
filter
No. of
persons
in
database
No. of
persons
out of
database
Number
of
images
of each
person
Number
of the
test
image
Optimum
threshold
EER TSR
‘Haar’
or ‘db1’
7 3 12 17 0.431552 0.100000 0.900000
‘db2’ 7 3 12 17 0.439712 0.142857 0.857143
‘db3’ 7 3 12 17 0.443662 0.100000 0.900000
‘db5’ 7 3 12 17 0.444544 0.142857 0.857143
‘db10’ 7 3 12 17 0.461352 0.100000 0.900000
'coif1 7 3 12 17 0.443977 0.142857 0.857143
'sym2' 7 3 12 17 0.439712 0.142857 0.857143
'dmey' 7 3 12 17 0.770870 0.000000 1.000000
31. 30
These are the results obtained when different bands such as LL, LH, HL
and HH bands are considered for feature extraction and matching. It has
been observed that better results are obtained for LL i.e. low-low band
which is generally used and good results for HL i.e. High-Low band.
The above plotted table is by considering HL band
No. of
persons
in
database
No. of
persons
out of
database
Number
of
images
of each
person
Number
of the
test
image
EER TSR
5 5 15 17 0.000000 1.000000
5 5 15 18 0.100000 0.900000
5 5 15 19 0.200000 0.800000
5 5 15 20 0.200000 0.800000
7 3 15 17 0.000000 1.000000
7 3 15 18 0.142857 0.857143
7 3 15 19 0.333333 0.666667
7 3 15 20 0.142857 0.857143
6 4 15 17 0.000000 1.000000
6 4 15 18 0.000000 1.000000
6 4 15 19 0.250000 0.750000
6 4 15 20 0.166667 0.833333
7 3 12 20 0.285714 0.714286
7 3 12 13 0.100000 0.900000
7 3 12 15 0.200000 0.800000
7 3 12 16 0.142857 0.857143
7 3 12 18 0.142857 0.857143
7 3 12 17 0.000000 1.000000
7 3 12 19 0.333333 0.666667