SlideShare a Scribd company logo
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 153
Face Recognition using Neural Networks
P.Latha plathamuthuraj@gmail.com
Selection .grade Lecturer,
Department of Electrical and Electronics Engineering,
Government College of Engineering,
Tirunelveli- 627007
Dr.L.Ganesan
Assistant Professor,
Head of Computer Science & Engineering department,
Alagappa Chettiar College of Engineering & Technology,
Karaikudi- 630004
Dr.S.Annadurai
Additional Director, Directorate of Technical Education
Chennai-600025
Abstract
Face recognition is one of biometric methods, to identify given face image using
main features of face. In this paper, a neural based algorithm is presented, to
detect frontal views of faces. The dimensionality of face image is reduced by the
Principal component analysis (PCA) and the recognition is done by the Back
propagation Neural Network (BPNN). Here 200 face images from Yale database
is taken and some performance metrics like Acceptance ratio and Execution time
are calculated. Neural based Face recognition is robust and has better
performance of more than 90 % acceptance ratio.
Key words: Face recognition-Principal Component Analysis- Back Propagation Neural Network -
Acceptance ratio–Execution time
1. INTRODUCTION
A face recognition system [6] is a computer vision and it automatically identifies a human face
from database images. The face recognition problem is challenging as it needs to account for all
possible appearance variation caused by change in illumination, facial features, occlusions, etc.
This paper gives a Neural and PCA based algorithm for efficient and robust face recognition.
Holistic approach, feature-based approach and hybrid approach are some of the approaches for
face recognition. Here, a holistic approach is used in which the whole face region is taken into
account as input data. This is based on principal component-analysis (PCA) technique, which is
used to simplify a dataset into lower dimension while retaining the characteristics of dataset.
Pre-processing, Principal component analysis and Back Propagation Neural Algorithm
are the major implementations of this paper. Pre-processing is done for two purposes
(i) To reduce noise and possible convolute effects of interfering system,
(ii) To transform the image into a different space where classification may prove
easier by exploitation of certain features.
PCA is a common statistical technique for finding the patterns in high dimensional data’s [1].
Feature extraction, also called Dimensionality Reduction, is done by PCA for a three main
purposes like
i) To reduce dimension of the data to more tractable limits
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 154
ii) To capture salient class-specific features of the data,
iii) To eliminate redundancy.
Here recognition is performed by both PCA and Back propagation Neural Networks [3].
BPNN mathematically models the behavior of the feature vectors by appropriate descriptions and
then exploits the statistical behavior of the feature vectors to define decision regions
corresponding to different classes. Any new pattern can be classified depending on which
decision region it would be falling in. All these processes are implemented for Face Recognition,
based on the basic block diagram as shown in fig 1.
Fig. 1 Basic Block Diagram
The Algorithm for Face recognition using neural classifier is as follows:
a) Pre-processing stage –Images are made zero-mean and unit-variance.
b) Dimensionality Reduction stage: PCA - Input data is reduced to a lower dimension to facilitate
classification.
c) Classification stage - The reduced vectors from PCA are applied to train BPNN classifier to
obtain the recognized image.
In this paper, Section 2 describes about Principal component analysis, Section 3 explains
about Back Propagation Neural Networks, Section 4 demonstrates experimentation and results
and subsequent chapters give conclusion and future development.
2. PRINCIPAL COMPONENT ANALYSIS
Principal component analysis (PCA) [2] involves a mathematical procedure that transforms a
number of possibly correlated variables into a smaller number of uncorrelated variables called
principal components. PCA is a popular technique, to derive a set of features for both face
recognition.
Any particular face can be
(i) Economically represented along the eigen pictures coordinate space, and
(ii) Approximately reconstructed using a small collection of Eigen pictures
To do this, a face image is projected to several face templates called eigenfaces which can be
considered as a set of features that characterize the variation between face images. Once a set
of eigenfaces is computed, a face image can be approximately reconstructed using a weighted
combination of the eigenfaces. The projection weights form a feature vector for face
representation and recognition. When a new test image is given, the weights are computed by
projecting the image onto the eigen- face vectors. The classification is then carried out by
comparing the distances between the weight vectors of the test image and the images from the
database. Conversely, using all of the eigenfaces extracted from the original images, one can
reconstruct the original image from the eigenfaces so that it matches the original image exactly.
2.1 PCA Algorithm
The algorithm used for principal component analysis is as follows.
(i) Acquire an initial set of M face images (the training set) & Calculate the eigen-faces
from the training set, keeping only M' eigenfaces that correspond to the highest
eigenvalue.
(ii) Calculate the corresponding distribution in M'-dimensional weight space for each
known individual, and calculate a set of weights based on the input image
(iii) Classify the weight pattern as either a known person or as unknown, according to its
distance to the closest weight vector of a known person.
Pre-
processed
Input Image
Principal
Component
Analysis
(PCA)
Back
Propagation
Neural Network
(BPNN)
Classified
Output
Image
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 155
Let the training set of images be MΓΓΓ ,....., 21 . The average face of the set is defined
by
∑=
Γ=Ψ
M
n
n
M 1
1
-----------(1)
Each face differs from the average by vector
Ψ−Γ=Φ ii ------------(2)
The co- variance matrix is formed by
T
M
n
T
nn AA
M
C ..
1
1
=ΦΦ= ∑= -----------(3)
where the matrix ].,.....,,[ 21 MA ΦΦΦ=
This set of large vectors is then subject to principal component analysis, which seeks a
set of M orthonormal vectors Muu ....1 .
To obtain a weight vector Ω of contributions of
individual eigen-faces to a facial image Γ, the face image is transformed into its eigen-face
components projected onto the face space by a simple operation
)( Ψ−Γ= T
kk uω -----------(4)
For k=1,.., M', where M' ≤ M is the number of eigen-faces used for the recognition. The weights
form vector Ω = [ ',......,, 21 M
ωωω ] that describes the contribution of each Eigen-face in
representing the face image Γ, treating the eigen-faces as a basis set for face images.The
simplest method for determining which face provides the best description of an unknown input
facial image is to find the image k that minimizes the Euclidean distance kε .
=kε || )( kΩ−Ω ||
2
------------(5)
where kΩ is a weight vector describing the k
th
face from the training set. A face is classified as
belonging to person k when the ‘ kε ‘is below some chosen threshold εΘ otherwise, the face is
classified as unknown.
The algorithm functions by projecting face images onto a feature space that spans the
significant variations among known face images. The projection operation characterizes an
individual face by a weighted sum of eigenfaces features, so to recognize a particular face, it is
necessary only to compare these weights to those of known individuals. The input image is
matched to the subject from the training set whose feature vector is the closest within acceptable
thresholds.
Eigen faces have advantages over the other techniques available, such as speed and
efficiency. For the system to work well in PCA, the faces must be seen from a frontal view under
similar lighting.
3. NEURAL NETWORKS AND BACK PROPAGATION ALGORITHM
A successful face recognition methodology depends heavily on the particular choice of the
features used by the pattern classifier .The Back-Propagation is the best known and widely used
learning algorithm in training multilayer perceptrons (MLP) [5]. The MLP refer to the network
consisting of a set of sensory units (source nodes) that constitute the input layer, one or more
hidden layers of computation nodes, and an output layer of computation nodes. The input signal
propagates through the network in a forward direction, from left to right and on a layer-by-layer
basis.
Back propagation is a multi-layer feed forward, supervised learning network based on gradient
descent learning rule. This BPNN provides a computationally efficient method for changing the
weights in feed forward network, with differentiable activation function units, to learn a training set
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 156
of input-output data. Being a gradient descent method it minimizes the total squared error of the
output computed by the net. The aim is to train the network to achieve a balance between the
ability to respond correctly to the input patterns that are used for training and the ability to provide
good response to the input that are similar.
3.1 Back Propagation Neural Networks Algorithm
A typical back propagation network [4] with Multi-layer, feed-forward supervised learning is as
shown in the figure. 2. Here learning process in Back propagation requires pairs of input and
target vectors. The output vector ‘o ‘is compared with target vector’t ‘. In case of difference of ‘o’
and‘t‘vectors, the weights are adjusted to minimize the difference. Initially random weights and
thresholds are assigned to the network. These weights are updated every iteration in order to
minimize the mean square error between the output vector and the target vector.
Fig. 2 Basic Block of Back propagation neural network
Input for hidden layer is given by
∑=
=
n
z
mzzm wxnet
1
----------- (6)
The units of output vector of hidden layer after passing through the activation function are given
by
( )m
m
net
h
−+
=
exp1
1
------------ (7)
In same manner, input for output layer is given by
kz
m
z
zk whnet ∑=
=
1
------------ (8)
and the units of output vector of output layer are given by
( )k
k
net
o
−+
=
exp1
1
----------- (9)
For updating the weights, we need to calculate the error. This can be done by
( )∑=
−=
k
li
ii toE
2
2
1
---------- (10)
oi and ti represents the real output and target output at neuron i in the output layer respectively. If
the error is minimum than a predefined limit, training process will stop; otherwise weights need to
be updated. For weights between hidden layer and output layer, the change in weights is given by
jiij hw αδ=∆ ----------- (11)
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 157
whereα is a training rate coefficient that is restricted to the range [0.01,1.0], hajj is the output of
neuron j in the hidden layer, and δi can be obtained by
( ) ( )iiiii oloot −−=δ ----------- (12)
Similarly, the change of the weights between hidden layer and output layer, is given by
jHiij xw βδ=∆ ----------- (13)
where β is a training rate coefficient that is restricted to the range [0.01,1.0], xj is the output of
neuron j in the input layer, and δ Hi can be obtained by
( ) ij
k
j
jiiHi wxlx ∑=
−=
1
δδ ----------- (14)
xi is the output at neuron i in the input layer, and summation term represents the weighted sum of
all δ j values corresponding to neurons in output layer that obtained in equation. After calculating
the weight change in all layers, the weights can simply updated by
( ) ( ) ijijij woldwneww ∆+= ----------- (15)
This process is repeated, until the error reaches a minimum value
2.4.3 Selection of Training Parameters
For the efficient operation of the back propagation network it is necessary for the appropriate
selection of the parameters used for training.
Initial Weights
This initial weight will influence whether the net reaches a global or local minima of the error and
if so how rapidly it converges. To get the best result the initial weights are set to random numbers
between -1 and 1.
Training a Net
The motivation for applying back propagation net is to achieve a balance between memorization
and generalization; it is not necessarily advantageous to continue training until the error reaches
a minimum value. The weight adjustments are based on the training patterns. As along as error
the for validation decreases training continues. Whenever the error begins to increase, the net is
starting to memorize the training patterns. At this point training is terminated.
Number of Hidden Units
If the activation function can vary with the function, then it can be seen that a n-input, m-
output function requires at most 2n+1 hidden units. If more number of hidden layers are present,
then the calculation for the δ’s are repeated for each additional hidden layer present, summing all
the δ’s for units present in the previous layer that is fed into the current layer for which δ is being
calculated.
Learning rate
In BPN, the weight change is in a direction that is a combination of current gradient and the
previous gradient. A small learning rate is used to avoid major disruption of the direction of
learning when very unusual pair of training patterns is presented.
Various parameters assumed for this algorithm are as follows.
No.of Input unit = 1 feature matrix
Accuracy = 0.001
learning rate = 0.4
No.of epochs = 400
No. of hidden neurons = 70
No.of output unit = 1
Main advantage of this back propagation algorithm is that it can identify the given image as a face
image or non face image and then recognizes the given input image .Thus the back propagation
neural network classifies the input image as recognized image.
4. Experimentation and Results
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 158
In this paper for experimentation, 200 images from Yale database are taken and a sample of 20
face images is as shown in fig 3. One of the images as shown in fig 4a is taken as the Input
image. The mean image and reconstructed output image by PCA, is as shown in fig 4b and 4c.
In BPNN, a training set of 50 images is as shown in fig 5a and the Eigen faces and recognized
output image are as shown in fig 5b and 5c.
Fig 3. Sample Yale Database Images
4(a) 4(b) 4 (c)
Fig 4.(a) Input Image , (b)Mean Image , (c) Recognized Image by PCA method
5(a) 5(b) 5(c)
Fig 5 (a) Training set, (b) Eigen faces , (c) Recognized Image by BPNN method
Table 1 shows the comparison of acceptance ratio and execution time values for 40, 80,
120,160 and 200 images of Yale database. Graphical analysis of the same is as shown in fig 6.
No .of Acceptance ratio (%) Execution Time (Seconds)
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 159
Images
PCA PCA with BPNN PCA PCA with BPNN
40 92.4 96.5 38 36
60 90.6 94.3 46 43
120 87.9 92.8 55 50
160 85.7 90.2 67 58
200 83.5 87.1 74 67
Table 1 Comparison of acceptance ratio and execution time for Yale database images
Fig.6: comparison of Acceptance ratio and execution time
5. CONCLUSION
Face recognition has received substantial attention from researches in biometrics, pattern
recognition field and computer vision communities. In this paper, Face recognition using Eigen
faces has been shown to be accurate and fast. When BPNN technique is combined with PCA,
non linear face images can be recognized easily. Hence it is concluded that this method has the
acceptance ratio is more than 90 % and execution time of only few seconds. Face recognition
can be applied in Security measure at Air ports, Passport verification, Criminals list verification in
police department, Visa processing , Verification of Electoral identification and Card Security
measure at ATM’s..
6. REFERENCES
[1]. B.K.Gunturk,A.U.Batur, and Y.Altunbasak,(2003) “Eigenface-domain super-resolution for
face recognition,” IEEE Transactions of . Image Processing. vol.12, no.5.pp. 597-606.
Comparision of Execution Time
0
10
20
30
40
50
60
70
80
40 60 120 160 200
No of images
ExecutionTime(sec)
PCA PCA with BPNN
Comparision of Acceptance Ratio
75
80
85
90
95
100
40 60 120 160 200
No of Images
AcceptanceRatio(%)
PCA PCA with BPNN
P.Latha, Dr.L.Ganesan & Dr.S.Annadurai
Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 160
[2]. M.A.Turk and A.P.Petland, (1991) “Eigenfaces for Recognition,” Journal of Cognitive
Neuroscience. vol. 3, pp.71-86.
[3]. T.Yahagi and H.Takano,(1994) “Face Recognition using neural networks with multiple
combinations of categories,” International Journal of Electronics Information and
Communication Engineering., vol.J77-D-II, no.11, pp.2151-2159.
[4]. S.Lawrence, C.L.Giles, A.C.Tsoi, and A.d.Back, (1993) “IEEE Transactions of Neural
Networks. vol.8, no.1, pp.98-113.
[5]. C.M.Bishop,(1995) “Neural Networks for Pattern Recognition” London, U.K.:Oxford
University Press.
[6]. Kailash J. Karande Sanjay N. Talbar “Independent Component Analysis of Edge
Information for Face Recognition” International Journal of Image Processing Volume (3) :
Issue (3) pp: 120 -131.

More Related Content

What's hot

Cnn
CnnCnn
Notes from Coursera Deep Learning courses by Andrew Ng
Notes from Coursera Deep Learning courses by Andrew NgNotes from Coursera Deep Learning courses by Andrew Ng
Notes from Coursera Deep Learning courses by Andrew Ng
dataHacker. rs
 
FACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORKFACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORK
TITHI DAN
 
1.Introduction to deep learning
1.Introduction to deep learning1.Introduction to deep learning
1.Introduction to deep learning
KONGU ENGINEERING COLLEGE
 
Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)
Muhammad Haroon
 
Convolutional Neural Network
Convolutional Neural NetworkConvolutional Neural Network
Convolutional Neural Network
Vignesh Suresh
 
Cat and dog classification
Cat and dog classificationCat and dog classification
Cat and dog classification
omaraldabash
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
Gaurav Mittal
 
Daa chapter 1
Daa chapter 1Daa chapter 1
Daa chapter 1
B.Kirron Reddi
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
Ashray Bhandare
 
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSING
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSINGBRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSING
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSING
Dharshika Shreeganesh
 
Image Classification using deep learning
Image Classification using deep learning Image Classification using deep learning
Image Classification using deep learning
Asma-AH
 
Batch normalization presentation
Batch normalization presentationBatch normalization presentation
Batch normalization presentation
Owin Will
 
Deep Learning - Overview of my work II
Deep Learning - Overview of my work IIDeep Learning - Overview of my work II
Deep Learning - Overview of my work II
Mohamed Loey
 
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical Analysis
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical AnalysisBrain Tumor Segmentation using Enhanced U-Net Model with Empirical Analysis
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical Analysis
MD Abdullah Al Nasim
 
Understanding Autoencoder (Deep Learning Book, Chapter 14)
Understanding Autoencoder  (Deep Learning Book, Chapter 14)Understanding Autoencoder  (Deep Learning Book, Chapter 14)
Understanding Autoencoder (Deep Learning Book, Chapter 14)
Entrepreneur / Startup
 
Activation function
Activation functionActivation function
Activation function
Astha Jain
 
Introduction to Deep learning
Introduction to Deep learningIntroduction to Deep learning
Introduction to Deep learning
Massimiliano Ruocco
 
Face Detection techniques
Face Detection techniquesFace Detection techniques
Face Detection techniques
Abhineet Bhamra
 
Convolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep LearningConvolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep Learning
Mohamed Loey
 

What's hot (20)

Cnn
CnnCnn
Cnn
 
Notes from Coursera Deep Learning courses by Andrew Ng
Notes from Coursera Deep Learning courses by Andrew NgNotes from Coursera Deep Learning courses by Andrew Ng
Notes from Coursera Deep Learning courses by Andrew Ng
 
FACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORKFACE RECOGNITION USING NEURAL NETWORK
FACE RECOGNITION USING NEURAL NETWORK
 
1.Introduction to deep learning
1.Introduction to deep learning1.Introduction to deep learning
1.Introduction to deep learning
 
Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)
 
Convolutional Neural Network
Convolutional Neural NetworkConvolutional Neural Network
Convolutional Neural Network
 
Cat and dog classification
Cat and dog classificationCat and dog classification
Cat and dog classification
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
 
Daa chapter 1
Daa chapter 1Daa chapter 1
Daa chapter 1
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
 
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSING
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSINGBRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSING
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSING
 
Image Classification using deep learning
Image Classification using deep learning Image Classification using deep learning
Image Classification using deep learning
 
Batch normalization presentation
Batch normalization presentationBatch normalization presentation
Batch normalization presentation
 
Deep Learning - Overview of my work II
Deep Learning - Overview of my work IIDeep Learning - Overview of my work II
Deep Learning - Overview of my work II
 
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical Analysis
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical AnalysisBrain Tumor Segmentation using Enhanced U-Net Model with Empirical Analysis
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical Analysis
 
Understanding Autoencoder (Deep Learning Book, Chapter 14)
Understanding Autoencoder  (Deep Learning Book, Chapter 14)Understanding Autoencoder  (Deep Learning Book, Chapter 14)
Understanding Autoencoder (Deep Learning Book, Chapter 14)
 
Activation function
Activation functionActivation function
Activation function
 
Introduction to Deep learning
Introduction to Deep learningIntroduction to Deep learning
Introduction to Deep learning
 
Face Detection techniques
Face Detection techniquesFace Detection techniques
Face Detection techniques
 
Convolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep LearningConvolutional Neural Network Models - Deep Learning
Convolutional Neural Network Models - Deep Learning
 

Similar to Face Recognition Using Neural Networks

Dx25751756
Dx25751756Dx25751756
Dx25751756
IJERA Editor
 
Comparison on PCA ICA and LDA in Face Recognition
Comparison on PCA ICA and LDA in Face RecognitionComparison on PCA ICA and LDA in Face Recognition
Comparison on PCA ICA and LDA in Face Recognition
ijdmtaiir
 
Face Recognition Using Gabor features And PCA
Face Recognition Using Gabor features And PCAFace Recognition Using Gabor features And PCA
Face Recognition Using Gabor features And PCA
IOSR Journals
 
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...
csandit
 
Fr pca lda
Fr pca ldaFr pca lda
Fr pca lda
ultraraj
 
COMPRESSION BASED FACE RECOGNITION USING DWT AND SVM
COMPRESSION BASED FACE RECOGNITION USING DWT AND SVMCOMPRESSION BASED FACE RECOGNITION USING DWT AND SVM
COMPRESSION BASED FACE RECOGNITION USING DWT AND SVM
sipij
 
fuzzy LBP for face recognition ppt
fuzzy LBP for face recognition pptfuzzy LBP for face recognition ppt
fuzzy LBP for face recognition ppt
Abdullah Gubbi
 
H0334749
H0334749H0334749
H0334749
iosrjournals
 
Y34147151
Y34147151Y34147151
Y34147151
IJERA Editor
 
Real-Time Face Tracking with GPU Acceleration
Real-Time Face Tracking with GPU AccelerationReal-Time Face Tracking with GPU Acceleration
Real-Time Face Tracking with GPU Acceleration
QuEST Global (erstwhile NeST Software)
 
CUDA Accelerated Face Recognition
CUDA Accelerated Face RecognitionCUDA Accelerated Face Recognition
CUDA Accelerated Face Recognition
QuEST Global (erstwhile NeST Software)
 
Application of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysisApplication of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysis
IAEME Publication
 
Application of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysisApplication of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysis
IAEME Publication
 
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...
ijma
 
Implementation of Face Recognition in Cloud Vision Using Eigen Faces
Implementation of Face Recognition in Cloud Vision Using Eigen FacesImplementation of Face Recognition in Cloud Vision Using Eigen Faces
Implementation of Face Recognition in Cloud Vision Using Eigen Faces
IJERA Editor
 
Image reconstruction through compressive sampling matching pursuit and curvel...
Image reconstruction through compressive sampling matching pursuit and curvel...Image reconstruction through compressive sampling matching pursuit and curvel...
Image reconstruction through compressive sampling matching pursuit and curvel...
IJECEIAES
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
Editor IJARCET
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
Editor IJARCET
 
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
IRJET Journal
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX model
IJECEIAES
 

Similar to Face Recognition Using Neural Networks (20)

Dx25751756
Dx25751756Dx25751756
Dx25751756
 
Comparison on PCA ICA and LDA in Face Recognition
Comparison on PCA ICA and LDA in Face RecognitionComparison on PCA ICA and LDA in Face Recognition
Comparison on PCA ICA and LDA in Face Recognition
 
Face Recognition Using Gabor features And PCA
Face Recognition Using Gabor features And PCAFace Recognition Using Gabor features And PCA
Face Recognition Using Gabor features And PCA
 
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...
 
Fr pca lda
Fr pca ldaFr pca lda
Fr pca lda
 
COMPRESSION BASED FACE RECOGNITION USING DWT AND SVM
COMPRESSION BASED FACE RECOGNITION USING DWT AND SVMCOMPRESSION BASED FACE RECOGNITION USING DWT AND SVM
COMPRESSION BASED FACE RECOGNITION USING DWT AND SVM
 
fuzzy LBP for face recognition ppt
fuzzy LBP for face recognition pptfuzzy LBP for face recognition ppt
fuzzy LBP for face recognition ppt
 
H0334749
H0334749H0334749
H0334749
 
Y34147151
Y34147151Y34147151
Y34147151
 
Real-Time Face Tracking with GPU Acceleration
Real-Time Face Tracking with GPU AccelerationReal-Time Face Tracking with GPU Acceleration
Real-Time Face Tracking with GPU Acceleration
 
CUDA Accelerated Face Recognition
CUDA Accelerated Face RecognitionCUDA Accelerated Face Recognition
CUDA Accelerated Face Recognition
 
Application of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysisApplication of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysis
 
Application of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysisApplication of gaussian filter with principal component analysis
Application of gaussian filter with principal component analysis
 
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...
 
Implementation of Face Recognition in Cloud Vision Using Eigen Faces
Implementation of Face Recognition in Cloud Vision Using Eigen FacesImplementation of Face Recognition in Cloud Vision Using Eigen Faces
Implementation of Face Recognition in Cloud Vision Using Eigen Faces
 
Image reconstruction through compressive sampling matching pursuit and curvel...
Image reconstruction through compressive sampling matching pursuit and curvel...Image reconstruction through compressive sampling matching pursuit and curvel...
Image reconstruction through compressive sampling matching pursuit and curvel...
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
 
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX model
 

Recently uploaded

Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
EduSkills OECD
 
Standardized tool for Intelligence test.
Standardized tool for Intelligence test.Standardized tool for Intelligence test.
Standardized tool for Intelligence test.
deepaannamalai16
 
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptx
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxBeyond Degrees - Empowering the Workforce in the Context of Skills-First.pptx
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptx
EduSkills OECD
 
Chapter wise All Notes of First year Basic Civil Engineering.pptx
Chapter wise All Notes of First year Basic Civil Engineering.pptxChapter wise All Notes of First year Basic Civil Engineering.pptx
Chapter wise All Notes of First year Basic Civil Engineering.pptx
Denish Jangid
 
Juneteenth Freedom Day 2024 David Douglas School District
Juneteenth Freedom Day 2024 David Douglas School DistrictJuneteenth Freedom Day 2024 David Douglas School District
Juneteenth Freedom Day 2024 David Douglas School District
David Douglas School District
 
مصحف القراءات العشر أعد أحرف الخلاف سمير بسيوني.pdf
مصحف القراءات العشر   أعد أحرف الخلاف سمير بسيوني.pdfمصحف القراءات العشر   أعد أحرف الخلاف سمير بسيوني.pdf
مصحف القراءات العشر أعد أحرف الخلاف سمير بسيوني.pdf
سمير بسيوني
 
Benner "Expanding Pathways to Publishing Careers"
Benner "Expanding Pathways to Publishing Careers"Benner "Expanding Pathways to Publishing Careers"
Benner "Expanding Pathways to Publishing Careers"
National Information Standards Organization (NISO)
 
Stack Memory Organization of 8086 Microprocessor
Stack Memory Organization of 8086 MicroprocessorStack Memory Organization of 8086 Microprocessor
Stack Memory Organization of 8086 Microprocessor
JomonJoseph58
 
Temple of Asclepius in Thrace. Excavation results
Temple of Asclepius in Thrace. Excavation resultsTemple of Asclepius in Thrace. Excavation results
Temple of Asclepius in Thrace. Excavation results
Krassimira Luka
 
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...
Nguyen Thanh Tu Collection
 
Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"
Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"
Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"
National Information Standards Organization (NISO)
 
HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.
deepaannamalai16
 
Nutrition Inc FY 2024, 4 - Hour Training
Nutrition Inc FY 2024, 4 - Hour TrainingNutrition Inc FY 2024, 4 - Hour Training
Nutrition Inc FY 2024, 4 - Hour Training
melliereed
 
Haunted Houses by H W Longfellow for class 10
Haunted Houses by H W Longfellow for class 10Haunted Houses by H W Longfellow for class 10
Haunted Houses by H W Longfellow for class 10
nitinpv4ai
 
Wound healing PPT
Wound healing PPTWound healing PPT
Wound healing PPT
Jyoti Chand
 
How to Predict Vendor Bill Product in Odoo 17
How to Predict Vendor Bill Product in Odoo 17How to Predict Vendor Bill Product in Odoo 17
How to Predict Vendor Bill Product in Odoo 17
Celine George
 
Skimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S EliotSkimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S Eliot
nitinpv4ai
 
Bossa N’ Roll Records by Ismael Vazquez.
Bossa N’ Roll Records by Ismael Vazquez.Bossa N’ Roll Records by Ismael Vazquez.
Bossa N’ Roll Records by Ismael Vazquez.
IsmaelVazquez38
 
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...
indexPub
 
Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...
Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...
Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...
imrankhan141184
 

Recently uploaded (20)

Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
 
Standardized tool for Intelligence test.
Standardized tool for Intelligence test.Standardized tool for Intelligence test.
Standardized tool for Intelligence test.
 
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptx
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxBeyond Degrees - Empowering the Workforce in the Context of Skills-First.pptx
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptx
 
Chapter wise All Notes of First year Basic Civil Engineering.pptx
Chapter wise All Notes of First year Basic Civil Engineering.pptxChapter wise All Notes of First year Basic Civil Engineering.pptx
Chapter wise All Notes of First year Basic Civil Engineering.pptx
 
Juneteenth Freedom Day 2024 David Douglas School District
Juneteenth Freedom Day 2024 David Douglas School DistrictJuneteenth Freedom Day 2024 David Douglas School District
Juneteenth Freedom Day 2024 David Douglas School District
 
مصحف القراءات العشر أعد أحرف الخلاف سمير بسيوني.pdf
مصحف القراءات العشر   أعد أحرف الخلاف سمير بسيوني.pdfمصحف القراءات العشر   أعد أحرف الخلاف سمير بسيوني.pdf
مصحف القراءات العشر أعد أحرف الخلاف سمير بسيوني.pdf
 
Benner "Expanding Pathways to Publishing Careers"
Benner "Expanding Pathways to Publishing Careers"Benner "Expanding Pathways to Publishing Careers"
Benner "Expanding Pathways to Publishing Careers"
 
Stack Memory Organization of 8086 Microprocessor
Stack Memory Organization of 8086 MicroprocessorStack Memory Organization of 8086 Microprocessor
Stack Memory Organization of 8086 Microprocessor
 
Temple of Asclepius in Thrace. Excavation results
Temple of Asclepius in Thrace. Excavation resultsTemple of Asclepius in Thrace. Excavation results
Temple of Asclepius in Thrace. Excavation results
 
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 9 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2024-2025 - ...
 
Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"
Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"
Jemison, MacLaughlin, and Majumder "Broadening Pathways for Editors and Authors"
 
HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.
 
Nutrition Inc FY 2024, 4 - Hour Training
Nutrition Inc FY 2024, 4 - Hour TrainingNutrition Inc FY 2024, 4 - Hour Training
Nutrition Inc FY 2024, 4 - Hour Training
 
Haunted Houses by H W Longfellow for class 10
Haunted Houses by H W Longfellow for class 10Haunted Houses by H W Longfellow for class 10
Haunted Houses by H W Longfellow for class 10
 
Wound healing PPT
Wound healing PPTWound healing PPT
Wound healing PPT
 
How to Predict Vendor Bill Product in Odoo 17
How to Predict Vendor Bill Product in Odoo 17How to Predict Vendor Bill Product in Odoo 17
How to Predict Vendor Bill Product in Odoo 17
 
Skimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S EliotSkimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S Eliot
 
Bossa N’ Roll Records by Ismael Vazquez.
Bossa N’ Roll Records by Ismael Vazquez.Bossa N’ Roll Records by Ismael Vazquez.
Bossa N’ Roll Records by Ismael Vazquez.
 
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...
 
Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...
Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...
Traditional Musical Instruments of Arunachal Pradesh and Uttar Pradesh - RAYH...
 

Face Recognition Using Neural Networks

  • 1. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 153 Face Recognition using Neural Networks P.Latha plathamuthuraj@gmail.com Selection .grade Lecturer, Department of Electrical and Electronics Engineering, Government College of Engineering, Tirunelveli- 627007 Dr.L.Ganesan Assistant Professor, Head of Computer Science & Engineering department, Alagappa Chettiar College of Engineering & Technology, Karaikudi- 630004 Dr.S.Annadurai Additional Director, Directorate of Technical Education Chennai-600025 Abstract Face recognition is one of biometric methods, to identify given face image using main features of face. In this paper, a neural based algorithm is presented, to detect frontal views of faces. The dimensionality of face image is reduced by the Principal component analysis (PCA) and the recognition is done by the Back propagation Neural Network (BPNN). Here 200 face images from Yale database is taken and some performance metrics like Acceptance ratio and Execution time are calculated. Neural based Face recognition is robust and has better performance of more than 90 % acceptance ratio. Key words: Face recognition-Principal Component Analysis- Back Propagation Neural Network - Acceptance ratio–Execution time 1. INTRODUCTION A face recognition system [6] is a computer vision and it automatically identifies a human face from database images. The face recognition problem is challenging as it needs to account for all possible appearance variation caused by change in illumination, facial features, occlusions, etc. This paper gives a Neural and PCA based algorithm for efficient and robust face recognition. Holistic approach, feature-based approach and hybrid approach are some of the approaches for face recognition. Here, a holistic approach is used in which the whole face region is taken into account as input data. This is based on principal component-analysis (PCA) technique, which is used to simplify a dataset into lower dimension while retaining the characteristics of dataset. Pre-processing, Principal component analysis and Back Propagation Neural Algorithm are the major implementations of this paper. Pre-processing is done for two purposes (i) To reduce noise and possible convolute effects of interfering system, (ii) To transform the image into a different space where classification may prove easier by exploitation of certain features. PCA is a common statistical technique for finding the patterns in high dimensional data’s [1]. Feature extraction, also called Dimensionality Reduction, is done by PCA for a three main purposes like i) To reduce dimension of the data to more tractable limits
  • 2. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 154 ii) To capture salient class-specific features of the data, iii) To eliminate redundancy. Here recognition is performed by both PCA and Back propagation Neural Networks [3]. BPNN mathematically models the behavior of the feature vectors by appropriate descriptions and then exploits the statistical behavior of the feature vectors to define decision regions corresponding to different classes. Any new pattern can be classified depending on which decision region it would be falling in. All these processes are implemented for Face Recognition, based on the basic block diagram as shown in fig 1. Fig. 1 Basic Block Diagram The Algorithm for Face recognition using neural classifier is as follows: a) Pre-processing stage –Images are made zero-mean and unit-variance. b) Dimensionality Reduction stage: PCA - Input data is reduced to a lower dimension to facilitate classification. c) Classification stage - The reduced vectors from PCA are applied to train BPNN classifier to obtain the recognized image. In this paper, Section 2 describes about Principal component analysis, Section 3 explains about Back Propagation Neural Networks, Section 4 demonstrates experimentation and results and subsequent chapters give conclusion and future development. 2. PRINCIPAL COMPONENT ANALYSIS Principal component analysis (PCA) [2] involves a mathematical procedure that transforms a number of possibly correlated variables into a smaller number of uncorrelated variables called principal components. PCA is a popular technique, to derive a set of features for both face recognition. Any particular face can be (i) Economically represented along the eigen pictures coordinate space, and (ii) Approximately reconstructed using a small collection of Eigen pictures To do this, a face image is projected to several face templates called eigenfaces which can be considered as a set of features that characterize the variation between face images. Once a set of eigenfaces is computed, a face image can be approximately reconstructed using a weighted combination of the eigenfaces. The projection weights form a feature vector for face representation and recognition. When a new test image is given, the weights are computed by projecting the image onto the eigen- face vectors. The classification is then carried out by comparing the distances between the weight vectors of the test image and the images from the database. Conversely, using all of the eigenfaces extracted from the original images, one can reconstruct the original image from the eigenfaces so that it matches the original image exactly. 2.1 PCA Algorithm The algorithm used for principal component analysis is as follows. (i) Acquire an initial set of M face images (the training set) & Calculate the eigen-faces from the training set, keeping only M' eigenfaces that correspond to the highest eigenvalue. (ii) Calculate the corresponding distribution in M'-dimensional weight space for each known individual, and calculate a set of weights based on the input image (iii) Classify the weight pattern as either a known person or as unknown, according to its distance to the closest weight vector of a known person. Pre- processed Input Image Principal Component Analysis (PCA) Back Propagation Neural Network (BPNN) Classified Output Image
  • 3. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 155 Let the training set of images be MΓΓΓ ,....., 21 . The average face of the set is defined by ∑= Γ=Ψ M n n M 1 1 -----------(1) Each face differs from the average by vector Ψ−Γ=Φ ii ------------(2) The co- variance matrix is formed by T M n T nn AA M C .. 1 1 =ΦΦ= ∑= -----------(3) where the matrix ].,.....,,[ 21 MA ΦΦΦ= This set of large vectors is then subject to principal component analysis, which seeks a set of M orthonormal vectors Muu ....1 . To obtain a weight vector Ω of contributions of individual eigen-faces to a facial image Γ, the face image is transformed into its eigen-face components projected onto the face space by a simple operation )( Ψ−Γ= T kk uω -----------(4) For k=1,.., M', where M' ≤ M is the number of eigen-faces used for the recognition. The weights form vector Ω = [ ',......,, 21 M ωωω ] that describes the contribution of each Eigen-face in representing the face image Γ, treating the eigen-faces as a basis set for face images.The simplest method for determining which face provides the best description of an unknown input facial image is to find the image k that minimizes the Euclidean distance kε . =kε || )( kΩ−Ω || 2 ------------(5) where kΩ is a weight vector describing the k th face from the training set. A face is classified as belonging to person k when the ‘ kε ‘is below some chosen threshold εΘ otherwise, the face is classified as unknown. The algorithm functions by projecting face images onto a feature space that spans the significant variations among known face images. The projection operation characterizes an individual face by a weighted sum of eigenfaces features, so to recognize a particular face, it is necessary only to compare these weights to those of known individuals. The input image is matched to the subject from the training set whose feature vector is the closest within acceptable thresholds. Eigen faces have advantages over the other techniques available, such as speed and efficiency. For the system to work well in PCA, the faces must be seen from a frontal view under similar lighting. 3. NEURAL NETWORKS AND BACK PROPAGATION ALGORITHM A successful face recognition methodology depends heavily on the particular choice of the features used by the pattern classifier .The Back-Propagation is the best known and widely used learning algorithm in training multilayer perceptrons (MLP) [5]. The MLP refer to the network consisting of a set of sensory units (source nodes) that constitute the input layer, one or more hidden layers of computation nodes, and an output layer of computation nodes. The input signal propagates through the network in a forward direction, from left to right and on a layer-by-layer basis. Back propagation is a multi-layer feed forward, supervised learning network based on gradient descent learning rule. This BPNN provides a computationally efficient method for changing the weights in feed forward network, with differentiable activation function units, to learn a training set
  • 4. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 156 of input-output data. Being a gradient descent method it minimizes the total squared error of the output computed by the net. The aim is to train the network to achieve a balance between the ability to respond correctly to the input patterns that are used for training and the ability to provide good response to the input that are similar. 3.1 Back Propagation Neural Networks Algorithm A typical back propagation network [4] with Multi-layer, feed-forward supervised learning is as shown in the figure. 2. Here learning process in Back propagation requires pairs of input and target vectors. The output vector ‘o ‘is compared with target vector’t ‘. In case of difference of ‘o’ and‘t‘vectors, the weights are adjusted to minimize the difference. Initially random weights and thresholds are assigned to the network. These weights are updated every iteration in order to minimize the mean square error between the output vector and the target vector. Fig. 2 Basic Block of Back propagation neural network Input for hidden layer is given by ∑= = n z mzzm wxnet 1 ----------- (6) The units of output vector of hidden layer after passing through the activation function are given by ( )m m net h −+ = exp1 1 ------------ (7) In same manner, input for output layer is given by kz m z zk whnet ∑= = 1 ------------ (8) and the units of output vector of output layer are given by ( )k k net o −+ = exp1 1 ----------- (9) For updating the weights, we need to calculate the error. This can be done by ( )∑= −= k li ii toE 2 2 1 ---------- (10) oi and ti represents the real output and target output at neuron i in the output layer respectively. If the error is minimum than a predefined limit, training process will stop; otherwise weights need to be updated. For weights between hidden layer and output layer, the change in weights is given by jiij hw αδ=∆ ----------- (11)
  • 5. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 157 whereα is a training rate coefficient that is restricted to the range [0.01,1.0], hajj is the output of neuron j in the hidden layer, and δi can be obtained by ( ) ( )iiiii oloot −−=δ ----------- (12) Similarly, the change of the weights between hidden layer and output layer, is given by jHiij xw βδ=∆ ----------- (13) where β is a training rate coefficient that is restricted to the range [0.01,1.0], xj is the output of neuron j in the input layer, and δ Hi can be obtained by ( ) ij k j jiiHi wxlx ∑= −= 1 δδ ----------- (14) xi is the output at neuron i in the input layer, and summation term represents the weighted sum of all δ j values corresponding to neurons in output layer that obtained in equation. After calculating the weight change in all layers, the weights can simply updated by ( ) ( ) ijijij woldwneww ∆+= ----------- (15) This process is repeated, until the error reaches a minimum value 2.4.3 Selection of Training Parameters For the efficient operation of the back propagation network it is necessary for the appropriate selection of the parameters used for training. Initial Weights This initial weight will influence whether the net reaches a global or local minima of the error and if so how rapidly it converges. To get the best result the initial weights are set to random numbers between -1 and 1. Training a Net The motivation for applying back propagation net is to achieve a balance between memorization and generalization; it is not necessarily advantageous to continue training until the error reaches a minimum value. The weight adjustments are based on the training patterns. As along as error the for validation decreases training continues. Whenever the error begins to increase, the net is starting to memorize the training patterns. At this point training is terminated. Number of Hidden Units If the activation function can vary with the function, then it can be seen that a n-input, m- output function requires at most 2n+1 hidden units. If more number of hidden layers are present, then the calculation for the δ’s are repeated for each additional hidden layer present, summing all the δ’s for units present in the previous layer that is fed into the current layer for which δ is being calculated. Learning rate In BPN, the weight change is in a direction that is a combination of current gradient and the previous gradient. A small learning rate is used to avoid major disruption of the direction of learning when very unusual pair of training patterns is presented. Various parameters assumed for this algorithm are as follows. No.of Input unit = 1 feature matrix Accuracy = 0.001 learning rate = 0.4 No.of epochs = 400 No. of hidden neurons = 70 No.of output unit = 1 Main advantage of this back propagation algorithm is that it can identify the given image as a face image or non face image and then recognizes the given input image .Thus the back propagation neural network classifies the input image as recognized image. 4. Experimentation and Results
  • 6. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 158 In this paper for experimentation, 200 images from Yale database are taken and a sample of 20 face images is as shown in fig 3. One of the images as shown in fig 4a is taken as the Input image. The mean image and reconstructed output image by PCA, is as shown in fig 4b and 4c. In BPNN, a training set of 50 images is as shown in fig 5a and the Eigen faces and recognized output image are as shown in fig 5b and 5c. Fig 3. Sample Yale Database Images 4(a) 4(b) 4 (c) Fig 4.(a) Input Image , (b)Mean Image , (c) Recognized Image by PCA method 5(a) 5(b) 5(c) Fig 5 (a) Training set, (b) Eigen faces , (c) Recognized Image by BPNN method Table 1 shows the comparison of acceptance ratio and execution time values for 40, 80, 120,160 and 200 images of Yale database. Graphical analysis of the same is as shown in fig 6. No .of Acceptance ratio (%) Execution Time (Seconds)
  • 7. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 159 Images PCA PCA with BPNN PCA PCA with BPNN 40 92.4 96.5 38 36 60 90.6 94.3 46 43 120 87.9 92.8 55 50 160 85.7 90.2 67 58 200 83.5 87.1 74 67 Table 1 Comparison of acceptance ratio and execution time for Yale database images Fig.6: comparison of Acceptance ratio and execution time 5. CONCLUSION Face recognition has received substantial attention from researches in biometrics, pattern recognition field and computer vision communities. In this paper, Face recognition using Eigen faces has been shown to be accurate and fast. When BPNN technique is combined with PCA, non linear face images can be recognized easily. Hence it is concluded that this method has the acceptance ratio is more than 90 % and execution time of only few seconds. Face recognition can be applied in Security measure at Air ports, Passport verification, Criminals list verification in police department, Visa processing , Verification of Electoral identification and Card Security measure at ATM’s.. 6. REFERENCES [1]. B.K.Gunturk,A.U.Batur, and Y.Altunbasak,(2003) “Eigenface-domain super-resolution for face recognition,” IEEE Transactions of . Image Processing. vol.12, no.5.pp. 597-606. Comparision of Execution Time 0 10 20 30 40 50 60 70 80 40 60 120 160 200 No of images ExecutionTime(sec) PCA PCA with BPNN Comparision of Acceptance Ratio 75 80 85 90 95 100 40 60 120 160 200 No of Images AcceptanceRatio(%) PCA PCA with BPNN
  • 8. P.Latha, Dr.L.Ganesan & Dr.S.Annadurai Signal Processing: An International Journal (SPIJ) Volume (3) : Issue (5) 160 [2]. M.A.Turk and A.P.Petland, (1991) “Eigenfaces for Recognition,” Journal of Cognitive Neuroscience. vol. 3, pp.71-86. [3]. T.Yahagi and H.Takano,(1994) “Face Recognition using neural networks with multiple combinations of categories,” International Journal of Electronics Information and Communication Engineering., vol.J77-D-II, no.11, pp.2151-2159. [4]. S.Lawrence, C.L.Giles, A.C.Tsoi, and A.d.Back, (1993) “IEEE Transactions of Neural Networks. vol.8, no.1, pp.98-113. [5]. C.M.Bishop,(1995) “Neural Networks for Pattern Recognition” London, U.K.:Oxford University Press. [6]. Kailash J. Karande Sanjay N. Talbar “Independent Component Analysis of Edge Information for Face Recognition” International Journal of Image Processing Volume (3) : Issue (3) pp: 120 -131.