SlideShare a Scribd company logo
1 of 57
CS 312 Character and
Pattern Recognition – Spring 2009
Dimensionality Reduction Using PCA/LDA
Chapter 3 (Duda et al.) – Section 3.8
Thanks to Prof. Bebis
Case Studies:
Face Recognition Using Dimensionality Reduction
M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, 3(1),
pp. 71-86, 1991.
D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE
Transactions on Pattern Analysis and Machine Intelligence, 18(8), pp. 831-836, 1996.
A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 23, no. 2, pp. 228-233, 2001.
2
Dimensionality Reduction
• One approach to deal with high dimensional data is by reducing
their dimensionality.
• Project high dimensional data onto a lower dimensional sub-
space using linear or non-linear transformations.
3
Dimensionality Reduction
• Linear transformations are simple to compute and tractable.
• Classical –linear- approaches:
– Principal Component Analysis (PCA)
– Fisher Discriminant Analysis (FDA)
( )t
i i iY U X b u a= =
k x 1 k x d d x 1 (k<<d)k x 1 k x d d x 1 (k<<d)
4
Principal Component Analysis (PCA)
• Each dimensionality reduction technique finds an
appropriate transformation by satisfying certain criteria
(e.g., information loss, data discrimination, etc.)
• The goal of PCA is to reduce the dimensionality of the data
while retaining as much as possible of the variation
present in the dataset.
5
Principal Component Analysis (PCA)
1 1 2 2
1 2
ˆ ...
where , ,..., isa basein the -dimensionalsub-space(K<N)
K K
K
x b u b u b u
u u u K
= + + +
ˆx x=
1 1 2 2
1 2
...
where , ,..., isa basein theoriginal N-dimensionalspace
N N
n
x a v a v a v
v v v
= + + +
• Find a basis in a low dimensional sub-space:
− Approximate vectors by projecting them in a low dimensional
sub-space:
(1) Original space representation:
(2) Lower-dimensional sub-space representation:
• Note: if K=N, then
6
Principal Component Analysis (PCA)
• Example (K=N):
7
Principal Component Analysis (PCA)
• Information loss
− Dimensionality reduction implies information loss !!
− PCA preserves as much information as possible:
• What is the “best” lower dimensional sub-space?
The “best” low-dimensional space is centered at the sample mean
and has directions determined by the “best” eigenvectors of the
covariance matrix of the data x.
− By “best” eigenvectors we mean those corresponding to the largest
eigenvalues ( i.e., “principal components”).
− Since the covariance matrix is real and symmetric, these
eigenvectors are orthogonal and form a set of basis vectors.
(see pp. 114-117 in textbook for a proof)
ˆmin || || (reconstruction error)x x−
8
Principal Component Analysis (PCA)
• Methodology
− Suppose x1, x2, ..., xM are N x 1 vectors
9
Principal Component Analysis (PCA)
• Methodology – cont.
( )T
i ib u x x= −
10
Principal Component Analysis (PCA)
• Eigenvalue spectrum
λi
K
λN
11
Principal Component Analysis (PCA)
• Linear transformation implied by PCA
− The linear transformation RN
→ RK
that performs the dimensionality
reduction is:
12
Principal Component Analysis (PCA)
• Geometric interpretation
− PCA projects the data along the directions where the data varies the
most.
− These directions are determined by the eigenvectors of the
covariance matrix corresponding to the largest eigenvalues.
− The magnitude of the eigenvalues corresponds to the variance of
the data along the eigenvector directions.
13
Principal Component Analysis (PCA)
• How many principal components to keep?
− To choose K, you can use the following criterion:
14
Principal Component Analysis (PCA)
• What is the error due to dimensionality reduction?
• It can be shown that the average error due to
dimensionality reduction is equal to:
e
15
Principal Component Analysis (PCA)
• Standardization
− The principal components are dependent on the units used to
measure the original variables as well as on the range of values
they assume.
− We should always standardize the data prior to using PCA.
− A common standardization method is to transform all the data to
have zero mean and unit standard deviation:
16
Principal Component Analysis (PCA)
• Case Study: Eigenfaces for Face Detection/Recognition
− M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of
Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, 1991.
• Face Recognition
− The simplest approach is to think of it as a template matching
problem
− Problems arise when performing recognition in a high-
dimensional space.
− Significant improvements can be achieved by first mapping
the data into a lower dimensionality space.
− How to find this lower-dimensional space?
17
Principal Component Analysis (PCA)
• Main idea behind eigenfaces
ˆΦ −Ψ
−Ψ
average face
18
Principal Component Analysis (PCA)
• Computation of the eigenfaces
19
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
20
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
ui
21
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
22
Principal Component Analysis (PCA)
• Representing faces onto this basis
ˆ
iΦ
23
Principal Component Analysis (PCA)
• Eigenvalue spectrum : Defined as set of Eigenvalues.
• Spectrum is shown as Eigenvalues plotted in the graph
shown below:
λi
K
M
λN
24
Principal Component Analysis (PCA)
• Representing faces onto this basis – cont.
25
Principal Component Analysis (PCA)
• Face Recognition Using Eigenfaces
26
Principal Component Analysis (PCA)
• Face Recognition Using Eigenfaces – cont.
− The distance er is called distance within the face space (difs)
− Comment: we can use the common Euclidean distance to compute
er, however, it has been reported that the Mahalanobis distance
performs better:
27
Principal Component Analysis (PCA)
• Face Detection Using Eigenfaces
28
Principal Component Analysis (PCA)
• Face Detection Using Eigenfaces – cont.
29
Principal Component Analysis (PCA)
• Reconstruction of
faces and non-faces
30
Principal Component Analysis (PCA)
• Applications
− Face detection, tracking, and recognition
dffs
31
Principal Component Analysis (PCA)
• Problems in face recognition:
− Background (de-emphasize the outside of the face – e.g., by
multiplying the input image by a 2D Gaussian window centered on
the face)
− Lighting conditions (performance degrades with light changes)
− Scale (performance decreases quickly with changes to head size)
− multi-scale eigenspaces
− scale input image to multiple sizes
− Orientation (performance decreases but not as fast as with scale
changes)
− plane rotations can be handled
− out-of-plane rotations are more difficult to handle
32
Principal Component Analysis (PCA)
• Dataset
− 16 subjects
− 3 orientations, 3 sizes
− 3 lighting conditions, 6 resolutions (512x512 ... 16x16)
− Total number of images: 2,592
33
Principal Component Analysis (PCA)
• Experiment 1
− Used various sets of 16 images for training
− One image/person, taken under the same conditions
− Classify the rest images as one of the 16 individuals
− 7 eigenfaces were used
− No rejections (i.e., no threshold on difs)
− Performed a large number of experiments and averaged the results:
− 96% correct averaged over light variation
− 85% correct averaged over orientation variation
− 64% correct averaged over size variation
34
Principal Component Analysis (PCA)
• Experiment 2
− Considered rejections (i.e., by thresholding difs)
− There is a tradeoff between correct recognition and rejections.
− Adjusting the threshold to achieve 100% recognition accuracy
resulted in:
− 19% rejections while varying lighting
− 39% rejections while varying orientation
− 60% rejections while varying size
• Experiment 3
− Reconstruction using
partial information
35
Principal Component Analysis (PCA)
• PCA and classification
− PCA is not always an optimal dimensionality-reduction procedure
for classification purposes.
• Multiple classes and PCA
− Suppose there are C classes in the training data.
− PCA is based on the sample covariance which characterizes the
scatter of the entire data set, irrespective of class-membership.
− The projection axes chosen by PCA might not provide good
discrimination power.
36
Principal Component Analysis (PCA)
• PCA and classification (cont’d)
37
Linear Discriminant Analysis (LDA)
• What is the goal of LDA?
− Perform dimensionality reduction “while preserving as much of
the class discriminatory information as possible”.
− Seeks to find directions along which the classes are best separated.
− Takes into consideration the scatter within-classes but also the
scatter between-classes.
− More capable of distinguishing image variation due to identity from
variation due to other sources such as illumination and expression.
38
Linear Discriminant Analysis (LDA)
39
Linear Discriminant Analysis (LDA)
• Notation
(Sb has at most rank C-1)
(each sub-matrix has
rank 1 or less, i.e., outer
product of two vectors)
1 1
( )( )
iMC
T
w j i j i
i j
S x x
= =
= − −∑∑ μ μ
40
Linear Discriminant Analysis (LDA)
• Methodology
− LDA computes a transformation that maximizes the between-class
scatter while minimizing the within-class scatter:
| | | |
max max
| | | |
T
b b
T
w w
U S U S
U S U S
=
%
%
T
U=y x
products of eigenvalues !
projection matrix
,b wS S% % : scatter matrices of the projected data y
41
Linear Discriminant Analysis (LDA)
• Linear transformation implied by LDA
− The linear transformation is given by a matrix U whose columns are
the eigenvectors of the above problem (i.e., called Fisherfaces).
− The LDA solution is given by the eigenvectors of the generalized
eigenvector problem:
− Important: Since Sb has at most rank C-1, the max number of
eigenvectors with non-zero eigenvalues is C-1 (i.e., max
dimensionality of sub-space is C-1)
42
Linear Discriminant Analysis (LDA)
• Does Sw
-1
always exist?
− If Sw is non-singular, we can obtain a conventional eigenvalue
problem by writing:
− In practice, Sw is often singular since the data are image vectors
with large dimensionality while the size of the data set is much
smaller (M << N )
43
Linear Discriminant Analysis (LDA)
• Does Sw
-1
always exist? – cont.
− To alleviate this problem, we can use PCA first:
1) PCA is first applied to the data set to reduce its dimensionality.
2) LDA is then applied to find the most discriminative directions:
44
Linear Discriminant Analysis (LDA)
• Case Study: Using Discriminant Eigenfeatures for Image
Retrieval
− D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image
Retrieval", IEEE Transactions on Pattern Analysis and Machine
Intelligence, vol. 18, no. 8, pp. 831-836, 1996.
• Content-based image retrieval
− The application being studied here is query-by-example image
retrieval.
− The paper deals with the problem of selecting a good set of image
features for content-based image retrieval.
45
Linear Discriminant Analysis (LDA)
• Assumptions
− "Well-framed" images are required as input for training and query-
by-example test probes.
− Only a small variation in the size, position, and orientation of the
objects in the images is allowed.
46
Linear Discriminant Analysis (LDA)
• Some terminology
− Most Expressive Features (MEF): the features (projections)
obtained using PCA.
− Most Discriminating Features (MDF): the features (projections)
obtained using LDA.
• Numerical problems
− When computing the eigenvalues/eigenvectors of Sw
-1
SBuk = λkuk
numerically, the computations can be unstable since Sw
-1
SB is not
always symmetric.
− See paper for a way to find the eigenvalues/eigenvectors in a stable
way.
47
Linear Discriminant Analysis (LDA)
• Factors unrelated to classification
− MEF vectors show the tendency of PCA to capture major variations
in the training set such as lighting direction.
− MDF vectors discount those factors unrelated to classification.
48
Linear Discriminant Analysis (LDA)
• Clustering effect
1) Generate the set of MEFs/MDFs for each image in the training set.
2) Given an query image, compute its MEFs/MDFs using the same
procedure.
3) Find the k closest neighbors for retrieval (e.g., using Euclidean
distance).
• Methodology
49
Linear Discriminant Analysis (LDA)
• Experiments and results
− Face images
− A set of face images was used with 2 expressions, 3 lighting conditions.
− Testing was performed using a disjoint set of images:
• One image, randomly chosen, from each individual.
50
Linear Discriminant Analysis (LDA)
51
Linear Discriminant Analysis (LDA)
− Examples of correct search probes
52
Linear Discriminant Analysis (LDA)
− Example of a failed search probe
53
Linear Discriminant Analysis (LDA)
• Case Study: PCA versus LDA
− A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on
Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228-
233, 2001.
• Is LDA always better than PCA?
− There has been a tendency in the computer vision community to
prefer LDA over PCA.
− This is mainly because LDA deals directly with discrimination
between classes while PCA does not pay attention to the underlying
class structure.
− Main results of this study:
(1) When the training set is small, PCA can outperform LDA.
(2) When the number of samples is large and representative for each
class, LDA outperforms PCA.
54
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
55
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
LDA is not always better when training set is small
56
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
LDA outperforms PCA when training set is large
END
57

More Related Content

What's hot

Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...Universitat Politècnica de Catalunya
 
Class imbalance problem1
Class imbalance problem1Class imbalance problem1
Class imbalance problem1chs71
 
Edge Detection and Segmentation
Edge Detection and SegmentationEdge Detection and Segmentation
Edge Detection and SegmentationA B Shinde
 
multiple linear regression
multiple linear regressionmultiple linear regression
multiple linear regressionAkhilesh Joshi
 
Markov Chain Monte Carlo explained
Markov Chain Monte Carlo explainedMarkov Chain Monte Carlo explained
Markov Chain Monte Carlo explaineddariodigiuni
 
Color image processing Presentation
Color image processing PresentationColor image processing Presentation
Color image processing PresentationRevanth Chimmani
 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentMuhammad Rasel
 
pattern classification
pattern classificationpattern classification
pattern classificationRanjan Ganguli
 
Noise filtering
Noise filteringNoise filtering
Noise filteringAlaa Ahmed
 
Expectation Maximization and Gaussian Mixture Models
Expectation Maximization and Gaussian Mixture ModelsExpectation Maximization and Gaussian Mixture Models
Expectation Maximization and Gaussian Mixture Modelspetitegeek
 
Presentation on unsupervised learning
Presentation on unsupervised learning Presentation on unsupervised learning
Presentation on unsupervised learning ANKUSH PAL
 
Classification Based Machine Learning Algorithms
Classification Based Machine Learning AlgorithmsClassification Based Machine Learning Algorithms
Classification Based Machine Learning AlgorithmsMd. Main Uddin Rony
 
Linear models
Linear modelsLinear models
Linear modelsFAO
 
07 regularization
07 regularization07 regularization
07 regularizationRonald Teo
 
Lecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation MaximizationLecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation Maximizationbutest
 

What's hot (20)

Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
 
Logistic regression
Logistic regressionLogistic regression
Logistic regression
 
Decision Tree Learning
Decision Tree LearningDecision Tree Learning
Decision Tree Learning
 
Class imbalance problem1
Class imbalance problem1Class imbalance problem1
Class imbalance problem1
 
Edge Detection and Segmentation
Edge Detection and SegmentationEdge Detection and Segmentation
Edge Detection and Segmentation
 
Principal component analysis
Principal component analysisPrincipal component analysis
Principal component analysis
 
multiple linear regression
multiple linear regressionmultiple linear regression
multiple linear regression
 
Markov Chain Monte Carlo explained
Markov Chain Monte Carlo explainedMarkov Chain Monte Carlo explained
Markov Chain Monte Carlo explained
 
Color image processing Presentation
Color image processing PresentationColor image processing Presentation
Color image processing Presentation
 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descent
 
pattern classification
pattern classificationpattern classification
pattern classification
 
Noise filtering
Noise filteringNoise filtering
Noise filtering
 
Expectation Maximization and Gaussian Mixture Models
Expectation Maximization and Gaussian Mixture ModelsExpectation Maximization and Gaussian Mixture Models
Expectation Maximization and Gaussian Mixture Models
 
Hog
HogHog
Hog
 
Particle filter
Particle filterParticle filter
Particle filter
 
Presentation on unsupervised learning
Presentation on unsupervised learning Presentation on unsupervised learning
Presentation on unsupervised learning
 
Classification Based Machine Learning Algorithms
Classification Based Machine Learning AlgorithmsClassification Based Machine Learning Algorithms
Classification Based Machine Learning Algorithms
 
Linear models
Linear modelsLinear models
Linear models
 
07 regularization
07 regularization07 regularization
07 regularization
 
Lecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation MaximizationLecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation Maximization
 

Viewers also liked

Linear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its GeneralizationLinear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its Generalization일상 온
 
face recognition system
face recognition systemface recognition system
face recognition systemAnil Kumar
 
LDA Beginner's Tutorial
LDA Beginner's TutorialLDA Beginner's Tutorial
LDA Beginner's TutorialWayne Lee
 
Some contributions to PCA for time series
Some contributions to PCA for time seriesSome contributions to PCA for time series
Some contributions to PCA for time seriesIsabel Magalhães
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet AllocationMarco Righini
 
Topic Models, LDA and all that
Topic Models, LDA and all thatTopic Models, LDA and all that
Topic Models, LDA and all thatZhibo Xiao
 
LDA presentation
LDA presentationLDA presentation
LDA presentationMohit Gupta
 
Text Mining using LDA with Context
Text Mining using LDA with ContextText Mining using LDA with Context
Text Mining using LDA with ContextSteffen Staab
 
PCA Based Face Recognition System
PCA Based Face Recognition SystemPCA Based Face Recognition System
PCA Based Face Recognition SystemMd. Atiqur Rahman
 
Face recognition technology - BEST PPT
Face recognition technology - BEST PPTFace recognition technology - BEST PPT
Face recognition technology - BEST PPTSiddharth Modi
 
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vecword2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec👋 Christopher Moody
 
2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda2012 mdsp pr09 pca lda
2012 mdsp pr09 pca ldanozomuhamada
 
Big Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statisticsBig Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statisticsDario Buono
 

Viewers also liked (20)

PCA vs LDA
PCA vs LDAPCA vs LDA
PCA vs LDA
 
Face recognition using LDA
Face recognition using LDAFace recognition using LDA
Face recognition using LDA
 
Linear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its GeneralizationLinear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its Generalization
 
face recognition system
face recognition systemface recognition system
face recognition system
 
Lda
LdaLda
Lda
 
LDA Beginner's Tutorial
LDA Beginner's TutorialLDA Beginner's Tutorial
LDA Beginner's Tutorial
 
Pca ppt
Pca pptPca ppt
Pca ppt
 
Some contributions to PCA for time series
Some contributions to PCA for time seriesSome contributions to PCA for time series
Some contributions to PCA for time series
 
LDA
LDALDA
LDA
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet Allocation
 
Topic Models, LDA and all that
Topic Models, LDA and all thatTopic Models, LDA and all that
Topic Models, LDA and all that
 
LDA presentation
LDA presentationLDA presentation
LDA presentation
 
Text Mining using LDA with Context
Text Mining using LDA with ContextText Mining using LDA with Context
Text Mining using LDA with Context
 
PCA Based Face Recognition System
PCA Based Face Recognition SystemPCA Based Face Recognition System
PCA Based Face Recognition System
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component Analysis
 
07. PCA
07. PCA07. PCA
07. PCA
 
Face recognition technology - BEST PPT
Face recognition technology - BEST PPTFace recognition technology - BEST PPT
Face recognition technology - BEST PPT
 
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vecword2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
 
2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda
 
Big Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statisticsBig Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statistics
 

Similar to Understandig PCA and LDA

Fr pca lda
Fr pca ldaFr pca lda
Fr pca ldaultraraj
 
DimensionalityReduction.pptx
DimensionalityReduction.pptxDimensionalityReduction.pptx
DimensionalityReduction.pptx36rajneekant
 
Dimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxDimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxSivam Chinna
 
pcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxpcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxABINASHPADHY6
 
introduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxintroduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxPratik Gohel
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Mostafa G. M. Mostafa
 
Matrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsMatrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsYONG ZHENG
 
5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdfRahul926331
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptorswolf
 
Singular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxSingular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxrajalakshmi5921
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxEDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxrajalakshmi5921
 
The following ppt is about principal component analysis
The following ppt is about principal component analysisThe following ppt is about principal component analysis
The following ppt is about principal component analysisSushmit8
 
Computational Giants_nhom.pptx
Computational Giants_nhom.pptxComputational Giants_nhom.pptx
Computational Giants_nhom.pptxThAnhonc
 
Lect4 principal component analysis-I
Lect4 principal component analysis-ILect4 principal component analysis-I
Lect4 principal component analysis-Ihktripathy
 
Face Recognition Using Neural Networks
Face Recognition Using Neural NetworksFace Recognition Using Neural Networks
Face Recognition Using Neural NetworksCSCJournals
 
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAminaRepo
 

Similar to Understandig PCA and LDA (20)

Fr pca lda
Fr pca ldaFr pca lda
Fr pca lda
 
DimensionalityReduction.pptx
DimensionalityReduction.pptxDimensionalityReduction.pptx
DimensionalityReduction.pptx
 
Dimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxDimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptx
 
pcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxpcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptx
 
introduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxintroduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptx
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)
 
Matrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsMatrix Factorization In Recommender Systems
Matrix Factorization In Recommender Systems
 
5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
 
Singular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxSingular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptx
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxEDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptx
 
pca.ppt
pca.pptpca.ppt
pca.ppt
 
The following ppt is about principal component analysis
The following ppt is about principal component analysisThe following ppt is about principal component analysis
The following ppt is about principal component analysis
 
PCA Final.pptx
PCA Final.pptxPCA Final.pptx
PCA Final.pptx
 
Fa18_P2.pptx
Fa18_P2.pptxFa18_P2.pptx
Fa18_P2.pptx
 
Machine learning meetup
Machine learning meetupMachine learning meetup
Machine learning meetup
 
Computational Giants_nhom.pptx
Computational Giants_nhom.pptxComputational Giants_nhom.pptx
Computational Giants_nhom.pptx
 
Lect4 principal component analysis-I
Lect4 principal component analysis-ILect4 principal component analysis-I
Lect4 principal component analysis-I
 
Face Recognition Using Neural Networks
Face Recognition Using Neural NetworksFace Recognition Using Neural Networks
Face Recognition Using Neural Networks
 
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
 

More from Dr. Syed Hassan Amin

Greenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonGreenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonDr. Syed Hassan Amin
 
Multitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionMultitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionDr. Syed Hassan Amin
 
Thin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeThin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeDr. Syed Hassan Amin
 
Improving Code Quality Through Effective Review Process
Improving Code Quality Through Effective  Review ProcessImproving Code Quality Through Effective  Review Process
Improving Code Quality Through Effective Review ProcessDr. Syed Hassan Amin
 
Software Project Management Tips and Tricks
Software Project Management Tips and TricksSoftware Project Management Tips and Tricks
Software Project Management Tips and TricksDr. Syed Hassan Amin
 
Improving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesImproving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesDr. Syed Hassan Amin
 
Learning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsLearning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsDr. Syed Hassan Amin
 
Understanding and Managing Technical Debt
Understanding and Managing Technical DebtUnderstanding and Managing Technical Debt
Understanding and Managing Technical DebtDr. Syed Hassan Amin
 
An OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontAn OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontDr. Syed Hassan Amin
 

More from Dr. Syed Hassan Amin (12)

Greenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonGreenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparison
 
Introduction To Docker
Introduction To  DockerIntroduction To  Docker
Introduction To Docker
 
Laravel Unit Testing
Laravel Unit TestingLaravel Unit Testing
Laravel Unit Testing
 
Multitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionMultitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq Recognition
 
Agile Scrum Methodology
Agile Scrum MethodologyAgile Scrum Methodology
Agile Scrum Methodology
 
Thin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeThin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better Code
 
Improving Code Quality Through Effective Review Process
Improving Code Quality Through Effective  Review ProcessImproving Code Quality Through Effective  Review Process
Improving Code Quality Through Effective Review Process
 
Software Project Management Tips and Tricks
Software Project Management Tips and TricksSoftware Project Management Tips and Tricks
Software Project Management Tips and Tricks
 
Improving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesImproving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design Principles
 
Learning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsLearning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve Jobs
 
Understanding and Managing Technical Debt
Understanding and Managing Technical DebtUnderstanding and Managing Technical Debt
Understanding and Managing Technical Debt
 
An OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontAn OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq Font
 

Recently uploaded

Rithik Kumar Singh codealpha pythohn.pdf
Rithik Kumar Singh codealpha pythohn.pdfRithik Kumar Singh codealpha pythohn.pdf
Rithik Kumar Singh codealpha pythohn.pdfrahulyadav957181
 
Bank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis ProjectBank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis ProjectBoston Institute of Analytics
 
Real-Time AI Streaming - AI Max Princeton
Real-Time AI  Streaming - AI Max PrincetonReal-Time AI  Streaming - AI Max Princeton
Real-Time AI Streaming - AI Max PrincetonTimothy Spann
 
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...Thomas Poetter
 
Learn How Data Science Changes Our World
Learn How Data Science Changes Our WorldLearn How Data Science Changes Our World
Learn How Data Science Changes Our WorldEduminds Learning
 
Digital Marketing Plan, how digital marketing works
Digital Marketing Plan, how digital marketing worksDigital Marketing Plan, how digital marketing works
Digital Marketing Plan, how digital marketing worksdeepakthakur548787
 
Decoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis ProjectDecoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis ProjectBoston Institute of Analytics
 
modul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptxmodul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptxaleedritatuxx
 
IBEF report on the Insurance market in India
IBEF report on the Insurance market in IndiaIBEF report on the Insurance market in India
IBEF report on the Insurance market in IndiaManalVerma4
 
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Boston Institute of Analytics
 
Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...Seán Kennedy
 
SMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptxSMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptxHaritikaChhatwal1
 
why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...
why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...
why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...Jack Cole
 
What To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptxWhat To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptxSimranPal17
 
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptxThe Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptxTasha Penwell
 
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024Susanna-Assunta Sansone
 
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...Dr Arash Najmaei ( Phd., MBA, BSc)
 
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdfEnglish-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdfblazblazml
 
Decoding Movie Sentiments: Analyzing Reviews with Data Analysis model
Decoding Movie Sentiments: Analyzing Reviews with Data Analysis modelDecoding Movie Sentiments: Analyzing Reviews with Data Analysis model
Decoding Movie Sentiments: Analyzing Reviews with Data Analysis modelBoston Institute of Analytics
 

Recently uploaded (20)

Rithik Kumar Singh codealpha pythohn.pdf
Rithik Kumar Singh codealpha pythohn.pdfRithik Kumar Singh codealpha pythohn.pdf
Rithik Kumar Singh codealpha pythohn.pdf
 
Bank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis ProjectBank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis Project
 
Real-Time AI Streaming - AI Max Princeton
Real-Time AI  Streaming - AI Max PrincetonReal-Time AI  Streaming - AI Max Princeton
Real-Time AI Streaming - AI Max Princeton
 
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
 
Learn How Data Science Changes Our World
Learn How Data Science Changes Our WorldLearn How Data Science Changes Our World
Learn How Data Science Changes Our World
 
Digital Marketing Plan, how digital marketing works
Digital Marketing Plan, how digital marketing worksDigital Marketing Plan, how digital marketing works
Digital Marketing Plan, how digital marketing works
 
Decoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis ProjectDecoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis Project
 
modul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptxmodul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptx
 
IBEF report on the Insurance market in India
IBEF report on the Insurance market in IndiaIBEF report on the Insurance market in India
IBEF report on the Insurance market in India
 
Data Analysis Project: Stroke Prediction
Data Analysis Project: Stroke PredictionData Analysis Project: Stroke Prediction
Data Analysis Project: Stroke Prediction
 
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
 
Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...
 
SMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptxSMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptx
 
why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...
why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...
why-transparency-and-traceability-are-essential-for-sustainable-supply-chains...
 
What To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptxWhat To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptx
 
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptxThe Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
 
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
 
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
 
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdfEnglish-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
 
Decoding Movie Sentiments: Analyzing Reviews with Data Analysis model
Decoding Movie Sentiments: Analyzing Reviews with Data Analysis modelDecoding Movie Sentiments: Analyzing Reviews with Data Analysis model
Decoding Movie Sentiments: Analyzing Reviews with Data Analysis model
 

Understandig PCA and LDA

  • 1. CS 312 Character and Pattern Recognition – Spring 2009 Dimensionality Reduction Using PCA/LDA Chapter 3 (Duda et al.) – Section 3.8 Thanks to Prof. Bebis Case Studies: Face Recognition Using Dimensionality Reduction M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, 3(1), pp. 71-86, 1991. D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(8), pp. 831-836, 1996. A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228-233, 2001.
  • 2. 2 Dimensionality Reduction • One approach to deal with high dimensional data is by reducing their dimensionality. • Project high dimensional data onto a lower dimensional sub- space using linear or non-linear transformations.
  • 3. 3 Dimensionality Reduction • Linear transformations are simple to compute and tractable. • Classical –linear- approaches: – Principal Component Analysis (PCA) – Fisher Discriminant Analysis (FDA) ( )t i i iY U X b u a= = k x 1 k x d d x 1 (k<<d)k x 1 k x d d x 1 (k<<d)
  • 4. 4 Principal Component Analysis (PCA) • Each dimensionality reduction technique finds an appropriate transformation by satisfying certain criteria (e.g., information loss, data discrimination, etc.) • The goal of PCA is to reduce the dimensionality of the data while retaining as much as possible of the variation present in the dataset.
  • 5. 5 Principal Component Analysis (PCA) 1 1 2 2 1 2 ˆ ... where , ,..., isa basein the -dimensionalsub-space(K<N) K K K x b u b u b u u u u K = + + + ˆx x= 1 1 2 2 1 2 ... where , ,..., isa basein theoriginal N-dimensionalspace N N n x a v a v a v v v v = + + + • Find a basis in a low dimensional sub-space: − Approximate vectors by projecting them in a low dimensional sub-space: (1) Original space representation: (2) Lower-dimensional sub-space representation: • Note: if K=N, then
  • 6. 6 Principal Component Analysis (PCA) • Example (K=N):
  • 7. 7 Principal Component Analysis (PCA) • Information loss − Dimensionality reduction implies information loss !! − PCA preserves as much information as possible: • What is the “best” lower dimensional sub-space? The “best” low-dimensional space is centered at the sample mean and has directions determined by the “best” eigenvectors of the covariance matrix of the data x. − By “best” eigenvectors we mean those corresponding to the largest eigenvalues ( i.e., “principal components”). − Since the covariance matrix is real and symmetric, these eigenvectors are orthogonal and form a set of basis vectors. (see pp. 114-117 in textbook for a proof) ˆmin || || (reconstruction error)x x−
  • 8. 8 Principal Component Analysis (PCA) • Methodology − Suppose x1, x2, ..., xM are N x 1 vectors
  • 9. 9 Principal Component Analysis (PCA) • Methodology – cont. ( )T i ib u x x= −
  • 10. 10 Principal Component Analysis (PCA) • Eigenvalue spectrum λi K λN
  • 11. 11 Principal Component Analysis (PCA) • Linear transformation implied by PCA − The linear transformation RN → RK that performs the dimensionality reduction is:
  • 12. 12 Principal Component Analysis (PCA) • Geometric interpretation − PCA projects the data along the directions where the data varies the most. − These directions are determined by the eigenvectors of the covariance matrix corresponding to the largest eigenvalues. − The magnitude of the eigenvalues corresponds to the variance of the data along the eigenvector directions.
  • 13. 13 Principal Component Analysis (PCA) • How many principal components to keep? − To choose K, you can use the following criterion:
  • 14. 14 Principal Component Analysis (PCA) • What is the error due to dimensionality reduction? • It can be shown that the average error due to dimensionality reduction is equal to: e
  • 15. 15 Principal Component Analysis (PCA) • Standardization − The principal components are dependent on the units used to measure the original variables as well as on the range of values they assume. − We should always standardize the data prior to using PCA. − A common standardization method is to transform all the data to have zero mean and unit standard deviation:
  • 16. 16 Principal Component Analysis (PCA) • Case Study: Eigenfaces for Face Detection/Recognition − M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, 1991. • Face Recognition − The simplest approach is to think of it as a template matching problem − Problems arise when performing recognition in a high- dimensional space. − Significant improvements can be achieved by first mapping the data into a lower dimensionality space. − How to find this lower-dimensional space?
  • 17. 17 Principal Component Analysis (PCA) • Main idea behind eigenfaces ˆΦ −Ψ −Ψ average face
  • 18. 18 Principal Component Analysis (PCA) • Computation of the eigenfaces
  • 19. 19 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont.
  • 20. 20 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont. ui
  • 21. 21 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont.
  • 22. 22 Principal Component Analysis (PCA) • Representing faces onto this basis ˆ iΦ
  • 23. 23 Principal Component Analysis (PCA) • Eigenvalue spectrum : Defined as set of Eigenvalues. • Spectrum is shown as Eigenvalues plotted in the graph shown below: λi K M λN
  • 24. 24 Principal Component Analysis (PCA) • Representing faces onto this basis – cont.
  • 25. 25 Principal Component Analysis (PCA) • Face Recognition Using Eigenfaces
  • 26. 26 Principal Component Analysis (PCA) • Face Recognition Using Eigenfaces – cont. − The distance er is called distance within the face space (difs) − Comment: we can use the common Euclidean distance to compute er, however, it has been reported that the Mahalanobis distance performs better:
  • 27. 27 Principal Component Analysis (PCA) • Face Detection Using Eigenfaces
  • 28. 28 Principal Component Analysis (PCA) • Face Detection Using Eigenfaces – cont.
  • 29. 29 Principal Component Analysis (PCA) • Reconstruction of faces and non-faces
  • 30. 30 Principal Component Analysis (PCA) • Applications − Face detection, tracking, and recognition dffs
  • 31. 31 Principal Component Analysis (PCA) • Problems in face recognition: − Background (de-emphasize the outside of the face – e.g., by multiplying the input image by a 2D Gaussian window centered on the face) − Lighting conditions (performance degrades with light changes) − Scale (performance decreases quickly with changes to head size) − multi-scale eigenspaces − scale input image to multiple sizes − Orientation (performance decreases but not as fast as with scale changes) − plane rotations can be handled − out-of-plane rotations are more difficult to handle
  • 32. 32 Principal Component Analysis (PCA) • Dataset − 16 subjects − 3 orientations, 3 sizes − 3 lighting conditions, 6 resolutions (512x512 ... 16x16) − Total number of images: 2,592
  • 33. 33 Principal Component Analysis (PCA) • Experiment 1 − Used various sets of 16 images for training − One image/person, taken under the same conditions − Classify the rest images as one of the 16 individuals − 7 eigenfaces were used − No rejections (i.e., no threshold on difs) − Performed a large number of experiments and averaged the results: − 96% correct averaged over light variation − 85% correct averaged over orientation variation − 64% correct averaged over size variation
  • 34. 34 Principal Component Analysis (PCA) • Experiment 2 − Considered rejections (i.e., by thresholding difs) − There is a tradeoff between correct recognition and rejections. − Adjusting the threshold to achieve 100% recognition accuracy resulted in: − 19% rejections while varying lighting − 39% rejections while varying orientation − 60% rejections while varying size • Experiment 3 − Reconstruction using partial information
  • 35. 35 Principal Component Analysis (PCA) • PCA and classification − PCA is not always an optimal dimensionality-reduction procedure for classification purposes. • Multiple classes and PCA − Suppose there are C classes in the training data. − PCA is based on the sample covariance which characterizes the scatter of the entire data set, irrespective of class-membership. − The projection axes chosen by PCA might not provide good discrimination power.
  • 36. 36 Principal Component Analysis (PCA) • PCA and classification (cont’d)
  • 37. 37 Linear Discriminant Analysis (LDA) • What is the goal of LDA? − Perform dimensionality reduction “while preserving as much of the class discriminatory information as possible”. − Seeks to find directions along which the classes are best separated. − Takes into consideration the scatter within-classes but also the scatter between-classes. − More capable of distinguishing image variation due to identity from variation due to other sources such as illumination and expression.
  • 39. 39 Linear Discriminant Analysis (LDA) • Notation (Sb has at most rank C-1) (each sub-matrix has rank 1 or less, i.e., outer product of two vectors) 1 1 ( )( ) iMC T w j i j i i j S x x = = = − −∑∑ μ μ
  • 40. 40 Linear Discriminant Analysis (LDA) • Methodology − LDA computes a transformation that maximizes the between-class scatter while minimizing the within-class scatter: | | | | max max | | | | T b b T w w U S U S U S U S = % % T U=y x products of eigenvalues ! projection matrix ,b wS S% % : scatter matrices of the projected data y
  • 41. 41 Linear Discriminant Analysis (LDA) • Linear transformation implied by LDA − The linear transformation is given by a matrix U whose columns are the eigenvectors of the above problem (i.e., called Fisherfaces). − The LDA solution is given by the eigenvectors of the generalized eigenvector problem: − Important: Since Sb has at most rank C-1, the max number of eigenvectors with non-zero eigenvalues is C-1 (i.e., max dimensionality of sub-space is C-1)
  • 42. 42 Linear Discriminant Analysis (LDA) • Does Sw -1 always exist? − If Sw is non-singular, we can obtain a conventional eigenvalue problem by writing: − In practice, Sw is often singular since the data are image vectors with large dimensionality while the size of the data set is much smaller (M << N )
  • 43. 43 Linear Discriminant Analysis (LDA) • Does Sw -1 always exist? – cont. − To alleviate this problem, we can use PCA first: 1) PCA is first applied to the data set to reduce its dimensionality. 2) LDA is then applied to find the most discriminative directions:
  • 44. 44 Linear Discriminant Analysis (LDA) • Case Study: Using Discriminant Eigenfeatures for Image Retrieval − D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 18, no. 8, pp. 831-836, 1996. • Content-based image retrieval − The application being studied here is query-by-example image retrieval. − The paper deals with the problem of selecting a good set of image features for content-based image retrieval.
  • 45. 45 Linear Discriminant Analysis (LDA) • Assumptions − "Well-framed" images are required as input for training and query- by-example test probes. − Only a small variation in the size, position, and orientation of the objects in the images is allowed.
  • 46. 46 Linear Discriminant Analysis (LDA) • Some terminology − Most Expressive Features (MEF): the features (projections) obtained using PCA. − Most Discriminating Features (MDF): the features (projections) obtained using LDA. • Numerical problems − When computing the eigenvalues/eigenvectors of Sw -1 SBuk = λkuk numerically, the computations can be unstable since Sw -1 SB is not always symmetric. − See paper for a way to find the eigenvalues/eigenvectors in a stable way.
  • 47. 47 Linear Discriminant Analysis (LDA) • Factors unrelated to classification − MEF vectors show the tendency of PCA to capture major variations in the training set such as lighting direction. − MDF vectors discount those factors unrelated to classification.
  • 48. 48 Linear Discriminant Analysis (LDA) • Clustering effect 1) Generate the set of MEFs/MDFs for each image in the training set. 2) Given an query image, compute its MEFs/MDFs using the same procedure. 3) Find the k closest neighbors for retrieval (e.g., using Euclidean distance). • Methodology
  • 49. 49 Linear Discriminant Analysis (LDA) • Experiments and results − Face images − A set of face images was used with 2 expressions, 3 lighting conditions. − Testing was performed using a disjoint set of images: • One image, randomly chosen, from each individual.
  • 51. 51 Linear Discriminant Analysis (LDA) − Examples of correct search probes
  • 52. 52 Linear Discriminant Analysis (LDA) − Example of a failed search probe
  • 53. 53 Linear Discriminant Analysis (LDA) • Case Study: PCA versus LDA − A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228- 233, 2001. • Is LDA always better than PCA? − There has been a tendency in the computer vision community to prefer LDA over PCA. − This is mainly because LDA deals directly with discrimination between classes while PCA does not pay attention to the underlying class structure. − Main results of this study: (1) When the training set is small, PCA can outperform LDA. (2) When the number of samples is large and representative for each class, LDA outperforms PCA.
  • 54. 54 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont.
  • 55. 55 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont. LDA is not always better when training set is small
  • 56. 56 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont. LDA outperforms PCA when training set is large

Editor's Notes

  1. Good afternoon and thank you everyone for coming. My talk today will describe the research I performed at the IRIS at USC, the object of this work being to build a computational framework that addresses the problem of motion analysis and interpretation.