SlideShare a Scribd company logo
Eric Xing © Eric Xing @ CMU, 2006-2010 1
Machine Learning
Data visualization and
dimensionality reduction
Eric Xing
Lecture 7, August 13, 2010
Eric Xing © Eric Xing @ CMU, 2006-2010 2
Text document retrieval/labelling
 Represent each document by a high-dimensional vector in the
space of words
Eric Xing © Eric Xing @ CMU, 2006-2010 3
Image retrieval/labelling












=
nx
x
x
x

2
1
Eric Xing © Eric Xing @ CMU, 2006-2010 4
Dimensionality Bottlenecks
 Data dimension
 Sensor response variables X:
 1,000,000 samples of an EM/Acoustic field on each of N sensors
 10242 pixels of a projected image on a IR camera sensor
 N2 expansion factor to account for all pairwise correlations
 Information dimension
 Number of free parameters describing probability densities f(X) or f(S|X)
 For known statistical model: info dim = model dim
 For unknown model: info dim = dim of density approximation
 Parametric-model driven dimension reduction
 DR by sufficiency, DR by maximum likelihood
 Data-driven dimension reduction
 Manifold learning, structure discovery
Eric Xing © Eric Xing @ CMU, 2006-2010 5
Intuition: how does your brain
store these pictures?
Eric Xing © Eric Xing @ CMU, 2006-2010 6
Brain Representation
Eric Xing © Eric Xing @ CMU, 2006-2010 7
Brain Representation
 Every pixel?
 Or perceptually
meaningful structure?
 Up-down pose
 Left-right pose
 Lighting direction
So, your brain successfully reduced
the high-dimensional inputs to an
intrinsically 3-dimensional
manifold!
Eric Xing © Eric Xing @ CMU, 2006-2010 8
Manifold Embedding
(Metric) data geometry (Non-metric) information geometry
Domain
are i.i.d. samples from
Two Geometries to Consider
Eric Xing © Eric Xing @ CMU, 2006-2010 9
 Data-driven projection to lower dimensional subsapce
 Extract low-dim structure from high-dim data
 Data may lie on curved (but locally linear) subspace
[1] Josh .B. Tenenbaum, Vin de Silva, and John C. Langford “A Global Geometric Framework for
Nonlinear Dimensionality Reduction” Science, 22 Dec 2000.
[2] Jose Costa, Neal Patwari and Alfred O. Hero, “Distributed Weighted Multidimensional Scaling for
Node Localization in Sensor Networks”, IEEE/ACM Trans. Sensor Networks, to appear 2005.
[3] Misha Belkin and Partha Niyogi, “Laplacian eigenmaps for dimensionality reduction and data
representation,” Neural Computation, 2003.
Data-driven DR
Eric Xing © Eric Xing @ CMU, 2006-2010 10
What is a Manifold?
 A manifold is a topological space which is locally Euclidean.
 Represents a very useful and challenging unsupervised
learning problem.
 In general, any object which is nearly "flat" on small
scales is a manifold.
Eric Xing © Eric Xing @ CMU, 2006-2010 11
Manifold Learning
 Discover low dimensional structures (smooth manifold) for
data in high dimension.
 Linear Approaches
 Principal component analysis.
 Multi dimensional scaling.
 Non Linear Approaches
 Local Linear Embedding
 ISOMAP
 Laplacian Eigenmap.
Eric Xing © Eric Xing @ CMU, 2006-2010 12
Principal component analysis
 Areas of variance in data are where items can be best discriminated
and key underlying phenomena observed
 If two items or dimensions are highly correlated or dependent
 They are likely to represent highly related phenomena
 We want to combine related variables, and focus on uncorrelated or independent ones,
especially those along which the observations have high variance
 We look for the phenomena underlying the observed covariance/co-
dependence in a set of variables
 These phenomena are called “factors” or “principal components” or
“independent components,” depending on the methods used
 Factor analysis: based on variance/covariance/correlation
 Independent Component Analysis: based on independence
Eric Xing © Eric Xing @ CMU, 2006-2010 13
An example:
Eric Xing © Eric Xing @ CMU, 2006-2010 14
Principal Component Analysis
 The new variables/dimensions
 Are linear combinations of the original
ones
 Are uncorrelated with one another
 Orthogonal in original dimension space
 Capture as much of the original
variance in the data as possible
 Are called Principal Components
 Orthogonal directions of
greatest variance in data
 Projections along PC1
discriminate the data most
along any one axis
 First principal component is the direction of
greatest variability (covariance) in the data
 Second is the next orthogonal (uncorrelated)
direction of greatest variability
 So first remove all the variability along the first component, and then
find the next direction of greatest variability
 And so on …
Original Variable A
OriginalVariableB
PC 1
PC 2
Eric Xing © Eric Xing @ CMU, 2006-2010 15
Computing the Components
 Projection of vector x onto an axis (dimension) u is uTx
 Direction of greatest variability is that in which the average square of
the projection is greatest:
Maximize uTXXTu
s.t uTu = 1
Construct Langrangian uTXXTu – λuTu
Vector of partial derivatives set to zero
xxTu – λu = (xxT – λI) u = 0
As u ≠ 0 then u must be an eigenvector of XXT with eigenvalue λ
 λ is the principal eigenvalue of the correlation matrix C= XXT
 The eigenvalue denotes the amount of variability captured along that
dimension
Eric Xing © Eric Xing @ CMU, 2006-2010 16
Computing the Components
 Similarly for the next axis, etc.
 So, the new axes are the eigenvectors of the matrix of
correlations of the original variables, which captures the
similarities of the original variables based on how data
samples project to them
 Geometrically: centering followed by rotation
 Linear transformation
Eric Xing © Eric Xing @ CMU, 2006-2010 17
 For symmetric matrices, eigenvectors for distinct eigenvalues
are orthogonal
 All eigenvalues of a real symmetric matrix are real.
 All eigenvalues of a positive semidefinite matrix are non-
negative
ℜ∈⇒==− λλ T
SSand0if IS
0vSvifthen,0, ≥⇒=≥ℜ∈∀ λλSwww Tn
02121212121 =•⇒≠= vvvSv λλλ and,},{},{},{
Eigenvalues & Eigenvectors
Eric Xing © Eric Xing @ CMU, 2006-2010 18
 Let be a square matrix with m linearly
independent eigenvectors (a “non-defective” matrix)
 Theorem: Exists an eigen decomposition
(cf. matrix diagonalization theorem)
 Columns of U are eigenvectors of S
 Diagonal elements of are eigenvalues of
Eigen/diagonal Decomposition
diagonal
Unique
for
distinct
eigen-
values
Eric Xing © Eric Xing @ CMU, 2006-2010 19
PCs, Variance and Least-Squares
 The first PC retains the greatest amount of variation in the
sample
 The kth PC retains the kth greatest fraction of the variation in
the sample
 The kth largest eigenvalue of the correlation matrix C is the
variance in the sample along the kth PC
 The least-squares view: PCs are a series of linear least
squares fits to a sample, each orthogonal to all previous ones
Eric Xing © Eric Xing @ CMU, 2006-2010 20
0
5
10
15
20
25
PC1 PC2 PC3 PC4 PC5 PC6 PC7 PC8 PC9 PC10
Variance(%)
How Many PCs?
 For n original dimensions, sample covariance matrix is nxn, and has
up to n eigenvectors. So n PCs.
 Where does dimensionality reduction come from?
Can ignore the components of lesser significance.
You do lose some information, but if the eigenvalues are small, you don’t
lose much
 n dimensions in original data
 calculate n eigenvectors and eigenvalues
 choose only the first p eigenvectors, based on their eigenvalues
 final data set has only p dimensions
Eric Xing © Eric Xing @ CMU, 2006-2010 21
Application: querying text doc.
Eric Xing © Eric Xing @ CMU, 2006-2010 22
Within .40
threshold
K is the number of singular values used
Eric Xing © Eric Xing @ CMU, 2006-2010 23
Summary:
 Principle
 Linear projection method to reduce the number of parameters
 Transfer a set of correlated variables into a new set of uncorrelated variables
 Map the data into a space of lower dimensionality
 Form of unsupervised learning
 Properties
 It can be viewed as a rotation of the existing axes to new positions in the space defined by
original variables
 New axes are orthogonal and represent the directions with maximum variability
 Application: In many settings in pattern recognition and retrieval, we
have a feature-object matrix.
 For text, the terms are features and the docs are objects.
 Could be opinions and users …
 This matrix may be redundant in dimensionality.
 Can work with low-rank approximation.
 If entries are missing (e.g., users’ opinions), can recover if dimensionality is low.
Eric Xing © Eric Xing @ CMU, 2006-2010 24
Going beyond
 What is the essence of the C matrix?
 The elements in C captures some kind of affinity between a
pair of data points in the semantic space
 We can replace it with any reasonable affinity measure
 E.g., MDS
 E.g., the geodistance ISOMAP
( ) matrixdistance:
2
ijji xxD −=
Eric Xing © Eric Xing @ CMU, 2006-2010 25
 Constructing neighbourhood graph G
 For each pair of points in G, Computing shortest path
distances ---- geodesic distances.
 Use Dijkstra's or Floyd's algorithm
 Apply kernel PCA for C given by the centred matrix of squared
geodesic distances.
 Project test points onto principal components as in kernel
PCA.
Nonlinear DR – Isomap
[Josh. Tenenbaum, Vin de Silva, John langford 2000]
Eric Xing © Eric Xing @ CMU, 2006-2010
3D data
2D coord chart
Error vs. dimensionality of coordinate chart
“Swiss Roll” dataset
Eric Xing © Eric Xing @ CMU, 2006-2010 27
PCA, MD vs ISOMAP
 The residual variance of PCA (open triangles), MDS (open
circles), and Isomap
Eric Xing © Eric Xing @ CMU, 2006-2010 28
ISOMAP algorithm Pros/Cons
Advantages:
 Nonlinear
 Globally optimal
 Guarantee asymptotically to recover the true
dimensionality
Drawback:
 May not be stable, dependent on topology of data
 As N increases, pair wise distances provide better
approximations to geodesics, but cost more computation
Eric Xing © Eric Xing @ CMU, 2006-2010 29
Local Linear Embedding (a.k.a
LLE)
 LLE is based on simple geometric intuitions.
 Suppose the data consist of N real-valued
vectors Xi, each of dimensionality D.
 Each data point and its neighbors expected to lie on or
close to a locally linear patch of the manifold.
Eric Xing © Eric Xing @ CMU, 2006-2010 30
Steps in LLE algorithm
 Assign neighbors to each data point
 Compute the weights Wij that best linearly
reconstruct the data point from its neighbors, solving the
constrained least-squares problem.
 Compute the low-dimensional embedding vectors best
reconstructed by Wij.
iX

iY

Eric Xing © Eric Xing @ CMU, 2006-2010 31
Fit locally, Think Globally
From Nonlinear
Dimensionality
Reduction by
Locally Linear
Embedding
Sam T. Roweis and
Lawrence K. Saul
∑ ∑−=Φ
i j
jijYWYY 2
||)(

Eric Xing © Eric Xing @ CMU, 2006-2010 32
Super-Resolution Through
Neighbor Embedding [Yeung et al CVPR 2004]
Training Xsi
Training Ysi
Testing Xt
Testing Yt
?
Eric Xing © Eric Xing @ CMU, 2006-2010 33
Training Xsi
Training Ysi
High dimensional Manifold
Low dimensional Manifold
Intuition
 Patches of the image lie on a manifold
Eric Xing © Eric Xing @ CMU, 2006-2010 34
Algorithm
1. Get feature vectors for each low resolution training
patch.
2. For each test patch feature vector find K nearest
neighboring feature vectors of training patches.
3. Find optimum weights to express each test patch vector
as a weighted sum of its K nearest neighbor vectors.
4. Use these weights for reconstruction of that test patch in
high resolution.
Eric Xing © Eric Xing @ CMU, 2006-2010 35
Training Xsi
Training Ysi
Testing Xt
Testing Yt
Results
Eric Xing © Eric Xing @ CMU, 2006-2010 36
Summary:
 Principle
 Linear and nonlinear projection method to reduce the number of parameters
 Transfer a set of correlated variables into a new set of uncorrelated variables
 Map the data into a space of lower dimensionality
 Form of unsupervised learning
 Applications
 PCA and Latent semantic indexing for text mining
 Isomap and Nonparametric Models of Image Deformation
 LLE and Isomap Analysis of Spectra and Colour Images
 Image Spaces and Video Trajectories: Using Isomap to Explore Video Sequences
 Mining the structural knowledge of high-dimensional medical data using isomap
Isomap Webpage: http://isomap.stanford.edu/
Applying PCA and LDA:
Eigen-faces and Fisher-faces
L. Fei-Fei
Computer Science Dept.
Stanford University
Machine learning in computer vision
• Aug 13, Lecture 7: Dimensionality reduction,
Manifold learning
– Eigen- and Fisher- faces
– Applications to object representation
8/8/2010 38L. Fei-Fei, Dragon Star 2010, Stanford
The Space of Faces
• An image is a point in a high dimensional space
– An N x M image is a point in RNM
– We can define vectors in this space as we did in the 2D case
+=
[Thanks to Chuck Dyer, Steve Seitz, Nishino]
Key Idea
}ˆ{ P
RLx=χ• Images in the possible set are highly correlated.
• So, compress them to a low-dimensional subspace that
captures key appearance characteristics of the visual DOFs.
• EIGENFACES: [Turk and Pentland]
USE PCA!
Principal Component Analysis (PCA)
• PCA is used to determine the most representing
features among data points.
– It computes the p-dimensional subspace such that the
projection of the data points onto the subspace has the
largest variance among all p-dimensional subspaces.
Illustration of PCA
One projection PCA projection
x1
x2
1 2
3
4
5
6
x1
x2
1 2
3
4
5
6
X1’
Illustration of PCA
x1
x2
1st principal
component
2rd principal
component
Mathematical Formulation
Find a transformation, W,
m-dimensional n-dimensionalOrthonormal
Total scatter matrix:
Wopt corresponds to m eigen-
vectors of ST
Eigenfaces
• PCA extracts the eigenvectors of A
– Gives a set of vectors v1, v2, v3, ...
– Each one of these vectors is a direction in face space
• what do these look like?
Projecting onto the Eigenfaces
• The eigenfaces v1, ..., vK span the space of faces
– A face is converted to eigenface coordinates by
Algorithm
1. Align training images x1, x2, …, xN
2. Compute average face u = 1/N Σ xi
3. Compute the difference image φi = xi – u
Training
Note that each image is formulated into a long vector!
Algorithm
Testing
1. Projection in Eigenface
Projection ωi = W(X – u), W = {eigenfaces}
2. Compare projections
ST = 1/NΣ φi φi
T = BBT, B=[φ1, φ2 … φN]
4. Compute the covariance matrix (total scatter matrix)
5. Compute the eigenvectors of the covariance
matrix , W
Illustration of Eigenfaces
These are the first 4 eigenvectors from a training set of 400
images (ORL Face Database). They look like faces, hence
called Eigenface.
The visualization of eigenvectors:
Eigenfaces look somewhat like generic faces.
Eigenvalues
Only selecting the top P eigenfaces  reduces the dimensionality.
Fewer eigenfaces result in more information loss, and hence less
discrimination between faces.
Reconstruction and Errors
P = 4
P = 200
P = 400
Summary for PCA and Eigenface
• Non-iterative, globally optimal solution
• PCA projection is optimal for reconstruction
from a low dimensional basis, but may NOT be
optimal for discrimination…
Linear Discriminant Analysis (LDA)
• Using Linear Discriminant Analysis (LDA) or
Fisher’s Linear Discriminant (FLD)
• Eigenfaces attempt to maximise the scatter of the
training images in face space, while Fisherfaces
attempt to maximise the between class scatter,
while minimising the within class scatter.
Illustration of the Projection
Poor Projection Good Projection
x1
x2
x1
x2
 Using two classes as example:
Comparing with PCA
Variables
• N Sample images:
• c classes:
• Average of each class:
• Total average:
{ }Nxx ,,1 
{ }cχχ ,,1 
∑=
∈ ikx
k
i
i x
N χ
µ
1
∑=
=
N
k
kx
N 1
1
µ
Scatters
• Scatter of class i: ( )( )T
ik
x
iki xxS
ik
µµ
χ
−∑ −=
∈
∑=
=
c
i
iW SS
1
( )( )∑ −−=
=
c
i
T
iiiBS
1
µµµµχ
BWT SSS +=
• Within class scatter:
• Between class scatter:
• Total scatter:
Illustration
2S
1S
BS
21 SSSW +=
x1
x2
Mathematical Formulation (1)
After projection:
Between class scatter (of y’s):
Within class scatter (of y’s):
k
T
k xWy =
WSWS B
T
B =
~
WSWS W
T
W =
~
Mathematical Formulation (2)
• The desired projection:
WSW
WSW
S
S
W
W
T
B
T
W
B
opt
WW
maxarg~
~
maxarg ==
miwSwS iWiiB ,,1 == λ
• How is it found ?  Generalized Eigenvectors
Data dimension is much larger than the
number of samples
The matrix is singular:
Nn >>
( ) cNSRank W −≤WS
Fisherface (PCA+FLD)
• Project with FLD to space
• Project with PCA to space
1−c k
T
fldk zWy =
WWSWW
WWSWW
W
pcaW
T
pca
T
pcaB
T
pca
T
fld
W
maxarg=
cN − k
T
pcak xWz =
WSWW T
T
pca
W
maxarg=
Illustration of FisherFace
• Fisherface
Results: Eigenface vs. Fisherface (1)
• Variation in Facial Expression, Eyewear, and Lighting
• Input: 160 images of 16 people
• Train: 159 images
• Test: 1 image
With
glasses
Without
glasses
3 Lighting
conditions
5 expressions
Eigenface vs. Fisherface (2)
discussion
• Removing the first three principal
components results in better performance
under variable lighting conditions
• The Firsherface methods had error rates
lower than the Eigenface method for the
small datasets tested.
Manifold Learning for
Object Representation
L. Fei-Fei
Computer Science Dept.
Stanford University
Machine learning in computer vision
• Aug 13, Lecture 7: Dimensionality reduction, Manifold
learning
– Eigen- and Fisher- faces
– Applications to object representation
(slides courtesy to David Thompson)
8/8/2010 68L. Fei-Fei, Dragon Star 2010, Stanford
plenoptic function
manifolds in vision
appearance variation
manifolds in vision
images from hormel corp.
deformation
manifolds in vision
images from www.golfswingphotos.com
Find a low-D basis for
describing high-D data.
X ~= X' S.T.
dim(X') << dim(X)
uncovers the intrinsic
dimensionality
manifold learning
If we knew all pairwise distances…
Chicago Raleigh Boston Seattle S.F. Austin Orlando
Chicago 0
Raleigh 641 0
Boston 851 608 0
Seattle 1733 2363 2488 0
S.F. 1855 2406 2696 684 0
Austin 972 1167 1691 1764 1495 0
Orlando 994 520 1105 2565 2458 1015 0
Distances calculated with geobytes.com/CityDistanceTool
Multidimensional Scaling (MDS)
For n data points, and a distance matrix D,
Dij =
...we can construct a m-dimensional space to preserve
inter-point distances by using the top eigenvectors of D
scaled by their eigenvalues
j
i
MDS result in 2D
Actual plot of cities
Don’t know distances
Don’t know distnaces
1. data compression
2. “curse of dimensionality”
3. de-noising
4. visualization
5. reasonable distance metrics
why do manifold learning?
reasonable distance metrics
?
reasonable distance metrics
?
linear interpolation
reasonable distance metrics
?
manifold interpolation
Isomap for images
 Build a data graph G.
 Vertices: images
 (u,v) is an edge iff SSD(u,v) is small
 For any two images, we approximate the
distance between them with the “shortest path”
on G
Isomap
1. Build a sparse graph with K-nearest neighbors
Dg =
(distance matrix is
sparse)
Isomap
2. Infer other interpoint distances by finding shortest
paths on the graph (Dijkstra's
algorithm).
Dg =
Isomap
shortest-distance on a graph is easy to
compute
Isomap results: hands
- preserves global structure
- few free parameters
- sensitive to noise, noise edges
- computationally expensive (dense matrix
eigen-reduction)
Isomap: pro and con
Leakage problem
Find a mapping to preserve
local linear relationships
between neighbors
Locally Linear Embedding
Locally Linear Embedding
1. Find weight matrix W of linear
coefficients:
Enforce sum-to-one constraint.
LLE: Two key steps
2. Find projected vectors Y to minimize
reconstruction error
must solve for whole dataset
simultaneously
LLE: Two key steps
LLE: Result
preserves local
topology
PCA
LLE
LLE: Result
LLE: Result
LLE: Result
- no local minima, one free parameter
- incremental & fast
- simple linear algebra operations
- can distort global structure
LLE: pro and con

More Related Content

What's hot

Tda presentation
Tda presentationTda presentation
Tda presentation
HJ van Veen
 
Pca ankita dubey
Pca ankita dubeyPca ankita dubey
Pca ankita dubey
Ankita Dubey
 
Reproducibility and differential analysis with selfish
Reproducibility and differential analysis with selfishReproducibility and differential analysis with selfish
Reproducibility and differential analysis with selfish
tuxette
 
a paper reading of table recognition
a paper reading of table recognitiona paper reading of table recognition
a paper reading of table recognition
Ning Lu
 
QTML2021 UAP Quantum Feature Map
QTML2021 UAP Quantum Feature MapQTML2021 UAP Quantum Feature Map
QTML2021 UAP Quantum Feature Map
Ha Phuong
 
Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)
Ha Phuong
 
4 tracking objects of deformable shapes
4 tracking objects of deformable shapes4 tracking objects of deformable shapes
4 tracking objects of deformable shapesprjpublications
 
Bayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco Scutari
Bayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco ScutariBayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco Scutari
Bayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco Scutari
Bayes Nets meetup London
 
Summary of survey papers on deep learning method to 3D data
Summary of survey papers on deep learning method to 3D dataSummary of survey papers on deep learning method to 3D data
Summary of survey papers on deep learning method to 3D data
Arithmer Inc.
 
Self-Assembling Hyper-heuristics: a proof of concept
Self-Assembling Hyper-heuristics: a proof of conceptSelf-Assembling Hyper-heuristics: a proof of concept
Self-Assembling Hyper-heuristics: a proof of conceptGerman Terrazas
 
Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biology Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biology
tuxette
 
On mesh
On meshOn mesh
Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biologyKernel methods for data integration in systems biology
Kernel methods for data integration in systems biology
tuxette
 
Textural Feature Extraction of Natural Objects for Image Classification
Textural Feature Extraction of Natural Objects for Image ClassificationTextural Feature Extraction of Natural Objects for Image Classification
Textural Feature Extraction of Natural Objects for Image Classification
CSCJournals
 
Topological data analysis
Topological data analysisTopological data analysis
Topological data analysis
Sunghyon Kyeong
 
Interpolation of-geofield-parameters
Interpolation of-geofield-parametersInterpolation of-geofield-parameters
Interpolation of-geofield-parametersCemal Ardil
 
Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...
Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...
Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...
ijcisjournal
 
A survey on graph kernels
A survey on graph kernelsA survey on graph kernels
A survey on graph kernels
vincyy
 
Lecture 16 KL Transform in Image Processing
Lecture 16 KL Transform in Image ProcessingLecture 16 KL Transform in Image Processing
Lecture 16 KL Transform in Image Processing
VARUN KUMAR
 
Viii sem
Viii semViii sem
Viii sem
Lavesh Kaushik
 

What's hot (20)

Tda presentation
Tda presentationTda presentation
Tda presentation
 
Pca ankita dubey
Pca ankita dubeyPca ankita dubey
Pca ankita dubey
 
Reproducibility and differential analysis with selfish
Reproducibility and differential analysis with selfishReproducibility and differential analysis with selfish
Reproducibility and differential analysis with selfish
 
a paper reading of table recognition
a paper reading of table recognitiona paper reading of table recognition
a paper reading of table recognition
 
QTML2021 UAP Quantum Feature Map
QTML2021 UAP Quantum Feature MapQTML2021 UAP Quantum Feature Map
QTML2021 UAP Quantum Feature Map
 
Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)
 
4 tracking objects of deformable shapes
4 tracking objects of deformable shapes4 tracking objects of deformable shapes
4 tracking objects of deformable shapes
 
Bayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco Scutari
Bayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco ScutariBayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco Scutari
Bayes Nets Meetup Sept 29th 2016 - Bayesian Network Modelling by Marco Scutari
 
Summary of survey papers on deep learning method to 3D data
Summary of survey papers on deep learning method to 3D dataSummary of survey papers on deep learning method to 3D data
Summary of survey papers on deep learning method to 3D data
 
Self-Assembling Hyper-heuristics: a proof of concept
Self-Assembling Hyper-heuristics: a proof of conceptSelf-Assembling Hyper-heuristics: a proof of concept
Self-Assembling Hyper-heuristics: a proof of concept
 
Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biology Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biology
 
On mesh
On meshOn mesh
On mesh
 
Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biologyKernel methods for data integration in systems biology
Kernel methods for data integration in systems biology
 
Textural Feature Extraction of Natural Objects for Image Classification
Textural Feature Extraction of Natural Objects for Image ClassificationTextural Feature Extraction of Natural Objects for Image Classification
Textural Feature Extraction of Natural Objects for Image Classification
 
Topological data analysis
Topological data analysisTopological data analysis
Topological data analysis
 
Interpolation of-geofield-parameters
Interpolation of-geofield-parametersInterpolation of-geofield-parameters
Interpolation of-geofield-parameters
 
Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...
Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...
Content Based Image Retrieval Using Gray Level Co-Occurance Matrix with SVD a...
 
A survey on graph kernels
A survey on graph kernelsA survey on graph kernels
A survey on graph kernels
 
Lecture 16 KL Transform in Image Processing
Lecture 16 KL Transform in Image ProcessingLecture 16 KL Transform in Image Processing
Lecture 16 KL Transform in Image Processing
 
Viii sem
Viii semViii sem
Viii sem
 

Viewers also liked

Visualizing Threats: Network Visualization for Cyber Security
Visualizing Threats: Network Visualization for Cyber SecurityVisualizing Threats: Network Visualization for Cyber Security
Visualizing Threats: Network Visualization for Cyber Security
Cambridge Intelligence
 
Plotcon 2016 Visualization Talk by Alexandra Johnson
Plotcon 2016 Visualization Talk  by Alexandra JohnsonPlotcon 2016 Visualization Talk  by Alexandra Johnson
Plotcon 2016 Visualization Talk by Alexandra Johnson
SigOpt
 
PROTEUS H2020
PROTEUS H2020 PROTEUS H2020
PROTEUS H2020
Bonaventura Del Monte
 
Visualization and Theories of Learning in Education
Visualization and Theories of Learning in EducationVisualization and Theories of Learning in Education
Visualization and Theories of Learning in Education
Liz Dorland
 
Manifold learning
Manifold learningManifold learning
Manifold learning
Wei Yang
 
Real Time Machine Learning Visualization with Spark
Real Time Machine Learning Visualization with SparkReal Time Machine Learning Visualization with Spark
Real Time Machine Learning Visualization with Spark
DataWorks Summit/Hadoop Summit
 
Manifold Learning
Manifold LearningManifold Learning
Manifold Learning
Phong Vo
 
AI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't Changed
AI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't ChangedAI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't Changed
AI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't Changed
Raffael Marty
 
Visualizing the Model Selection Process
Visualizing the Model Selection ProcessVisualizing the Model Selection Process
Visualizing the Model Selection Process
Benjamin Bengfort
 
Dynamics in graph analysis (PyData Carolinas 2016)
Dynamics in graph analysis (PyData Carolinas 2016)Dynamics in graph analysis (PyData Carolinas 2016)
Dynamics in graph analysis (PyData Carolinas 2016)
Benjamin Bengfort
 

Viewers also liked (10)

Visualizing Threats: Network Visualization for Cyber Security
Visualizing Threats: Network Visualization for Cyber SecurityVisualizing Threats: Network Visualization for Cyber Security
Visualizing Threats: Network Visualization for Cyber Security
 
Plotcon 2016 Visualization Talk by Alexandra Johnson
Plotcon 2016 Visualization Talk  by Alexandra JohnsonPlotcon 2016 Visualization Talk  by Alexandra Johnson
Plotcon 2016 Visualization Talk by Alexandra Johnson
 
PROTEUS H2020
PROTEUS H2020 PROTEUS H2020
PROTEUS H2020
 
Visualization and Theories of Learning in Education
Visualization and Theories of Learning in EducationVisualization and Theories of Learning in Education
Visualization and Theories of Learning in Education
 
Manifold learning
Manifold learningManifold learning
Manifold learning
 
Real Time Machine Learning Visualization with Spark
Real Time Machine Learning Visualization with SparkReal Time Machine Learning Visualization with Spark
Real Time Machine Learning Visualization with Spark
 
Manifold Learning
Manifold LearningManifold Learning
Manifold Learning
 
AI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't Changed
AI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't ChangedAI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't Changed
AI & ML in Cyber Security - Welcome Back to 1999 - Security Hasn't Changed
 
Visualizing the Model Selection Process
Visualizing the Model Selection ProcessVisualizing the Model Selection Process
Visualizing the Model Selection Process
 
Dynamics in graph analysis (PyData Carolinas 2016)
Dynamics in graph analysis (PyData Carolinas 2016)Dynamics in graph analysis (PyData Carolinas 2016)
Dynamics in graph analysis (PyData Carolinas 2016)
 

Similar to Lecture7 xing fei-fei

Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...
Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...
Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...
Förderverein Technische Fakultät
 
Class9_PCA_final.ppt
Class9_PCA_final.pptClass9_PCA_final.ppt
Class9_PCA_final.ppt
MaTruongThanh002937
 
Lecture15 xing
Lecture15 xingLecture15 xing
Lecture15 xing
Tianlu Wang
 
Energy management system
Energy management systemEnergy management system
Energy management system
AmishaSrivastava26
 
Rigorous Pack Edge Detection Fuzzy System
Rigorous Pack Edge Detection Fuzzy SystemRigorous Pack Edge Detection Fuzzy System
Rigorous Pack Edge Detection Fuzzy System
inventy
 
BPSO&1-NN algorithm-based variable selection for power system stability ident...
BPSO&1-NN algorithm-based variable selection for power system stability ident...BPSO&1-NN algorithm-based variable selection for power system stability ident...
BPSO&1-NN algorithm-based variable selection for power system stability ident...
IJAEMSJORNAL
 
Nimrita koul Machine Learning
Nimrita koul  Machine LearningNimrita koul  Machine Learning
Nimrita koul Machine Learning
Nimrita Koul
 
4-RSSI-Spectral Domain Image Transforms_1.pdf
4-RSSI-Spectral Domain Image Transforms_1.pdf4-RSSI-Spectral Domain Image Transforms_1.pdf
4-RSSI-Spectral Domain Image Transforms_1.pdf
muhammadwalidmido
 
[Term project] Junction-point process
[Term project] Junction-point process[Term project] Junction-point process
[Term project] Junction-point process
Kyunghoon Kim
 
Multiscale Granger Causality and Information Decomposition
Multiscale Granger Causality and Information Decomposition Multiscale Granger Causality and Information Decomposition
Multiscale Granger Causality and Information Decomposition
danielemarinazzo
 
Introduction to Principle Component Analysis
Introduction to Principle Component AnalysisIntroduction to Principle Component Analysis
Introduction to Principle Component Analysis
Sunjeet Jena
 
Graphical Structure Learning accelerated with POWER9
Graphical Structure Learning accelerated with POWER9Graphical Structure Learning accelerated with POWER9
Graphical Structure Learning accelerated with POWER9
Ganesan Narayanasamy
 
Reduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theoryReduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theory
csandit
 
06-07 Chapter interpolation in MATLAB
06-07 Chapter interpolation in MATLAB06-07 Chapter interpolation in MATLAB
06-07 Chapter interpolation in MATLAB
Dr. Mohammed Danish
 
Fault diagnosis using genetic algorithms and principal curves
Fault diagnosis using genetic algorithms and principal curvesFault diagnosis using genetic algorithms and principal curves
Fault diagnosis using genetic algorithms and principal curves
eSAT Journals
 
Reverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturingReverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturing
AchmadRifaie4
 
3680-NoCA.pptx
3680-NoCA.pptx3680-NoCA.pptx
3680-NoCA.pptxgrssieee
 
Vol 9 No 1 - January 2014
Vol 9 No 1 - January 2014Vol 9 No 1 - January 2014
Vol 9 No 1 - January 2014
ijcsbi
 

Similar to Lecture7 xing fei-fei (20)

Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...
Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...
Nonequilibrium Network Dynamics_Inference, Fluctuation-Respones & Tipping Poi...
 
Class9_PCA_final.ppt
Class9_PCA_final.pptClass9_PCA_final.ppt
Class9_PCA_final.ppt
 
Lecture15 xing
Lecture15 xingLecture15 xing
Lecture15 xing
 
Energy management system
Energy management systemEnergy management system
Energy management system
 
Rigorous Pack Edge Detection Fuzzy System
Rigorous Pack Edge Detection Fuzzy SystemRigorous Pack Edge Detection Fuzzy System
Rigorous Pack Edge Detection Fuzzy System
 
BPSO&1-NN algorithm-based variable selection for power system stability ident...
BPSO&1-NN algorithm-based variable selection for power system stability ident...BPSO&1-NN algorithm-based variable selection for power system stability ident...
BPSO&1-NN algorithm-based variable selection for power system stability ident...
 
Nimrita koul Machine Learning
Nimrita koul  Machine LearningNimrita koul  Machine Learning
Nimrita koul Machine Learning
 
4-RSSI-Spectral Domain Image Transforms_1.pdf
4-RSSI-Spectral Domain Image Transforms_1.pdf4-RSSI-Spectral Domain Image Transforms_1.pdf
4-RSSI-Spectral Domain Image Transforms_1.pdf
 
[Term project] Junction-point process
[Term project] Junction-point process[Term project] Junction-point process
[Term project] Junction-point process
 
Multiscale Granger Causality and Information Decomposition
Multiscale Granger Causality and Information Decomposition Multiscale Granger Causality and Information Decomposition
Multiscale Granger Causality and Information Decomposition
 
Saif_CCECE2007_full_paper_submitted
Saif_CCECE2007_full_paper_submittedSaif_CCECE2007_full_paper_submitted
Saif_CCECE2007_full_paper_submitted
 
Introduction to Principle Component Analysis
Introduction to Principle Component AnalysisIntroduction to Principle Component Analysis
Introduction to Principle Component Analysis
 
Graphical Structure Learning accelerated with POWER9
Graphical Structure Learning accelerated with POWER9Graphical Structure Learning accelerated with POWER9
Graphical Structure Learning accelerated with POWER9
 
Reduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theoryReduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theory
 
call for papers, research paper publishing, where to publish research paper, ...
call for papers, research paper publishing, where to publish research paper, ...call for papers, research paper publishing, where to publish research paper, ...
call for papers, research paper publishing, where to publish research paper, ...
 
06-07 Chapter interpolation in MATLAB
06-07 Chapter interpolation in MATLAB06-07 Chapter interpolation in MATLAB
06-07 Chapter interpolation in MATLAB
 
Fault diagnosis using genetic algorithms and principal curves
Fault diagnosis using genetic algorithms and principal curvesFault diagnosis using genetic algorithms and principal curves
Fault diagnosis using genetic algorithms and principal curves
 
Reverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturingReverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturing
 
3680-NoCA.pptx
3680-NoCA.pptx3680-NoCA.pptx
3680-NoCA.pptx
 
Vol 9 No 1 - January 2014
Vol 9 No 1 - January 2014Vol 9 No 1 - January 2014
Vol 9 No 1 - January 2014
 

More from Tianlu Wang

14 pro resolution
14 pro resolution14 pro resolution
14 pro resolution
Tianlu Wang
 
13 propositional calculus
13 propositional calculus13 propositional calculus
13 propositional calculus
Tianlu Wang
 
12 adversal search
12 adversal search12 adversal search
12 adversal search
Tianlu Wang
 
11 alternative search
11 alternative search11 alternative search
11 alternative search
Tianlu Wang
 
21 situation calculus
21 situation calculus21 situation calculus
21 situation calculus
Tianlu Wang
 
20 bayes learning
20 bayes learning20 bayes learning
20 bayes learning
Tianlu Wang
 
19 uncertain evidence
19 uncertain evidence19 uncertain evidence
19 uncertain evidence
Tianlu Wang
 
18 common knowledge
18 common knowledge18 common knowledge
18 common knowledge
Tianlu Wang
 
17 2 expert systems
17 2 expert systems17 2 expert systems
17 2 expert systems
Tianlu Wang
 
17 1 knowledge-based system
17 1 knowledge-based system17 1 knowledge-based system
17 1 knowledge-based system
Tianlu Wang
 
16 2 predicate resolution
16 2 predicate resolution16 2 predicate resolution
16 2 predicate resolution
Tianlu Wang
 
16 1 predicate resolution
16 1 predicate resolution16 1 predicate resolution
16 1 predicate resolution
Tianlu Wang
 
09 heuristic search
09 heuristic search09 heuristic search
09 heuristic search
Tianlu Wang
 
08 uninformed search
08 uninformed search08 uninformed search
08 uninformed search
Tianlu Wang
 

More from Tianlu Wang (20)

L7 er2
L7 er2L7 er2
L7 er2
 
L8 design1
L8 design1L8 design1
L8 design1
 
L9 design2
L9 design2L9 design2
L9 design2
 
14 pro resolution
14 pro resolution14 pro resolution
14 pro resolution
 
13 propositional calculus
13 propositional calculus13 propositional calculus
13 propositional calculus
 
12 adversal search
12 adversal search12 adversal search
12 adversal search
 
11 alternative search
11 alternative search11 alternative search
11 alternative search
 
10 2 sum
10 2 sum10 2 sum
10 2 sum
 
22 planning
22 planning22 planning
22 planning
 
21 situation calculus
21 situation calculus21 situation calculus
21 situation calculus
 
20 bayes learning
20 bayes learning20 bayes learning
20 bayes learning
 
19 uncertain evidence
19 uncertain evidence19 uncertain evidence
19 uncertain evidence
 
18 common knowledge
18 common knowledge18 common knowledge
18 common knowledge
 
17 2 expert systems
17 2 expert systems17 2 expert systems
17 2 expert systems
 
17 1 knowledge-based system
17 1 knowledge-based system17 1 knowledge-based system
17 1 knowledge-based system
 
16 2 predicate resolution
16 2 predicate resolution16 2 predicate resolution
16 2 predicate resolution
 
16 1 predicate resolution
16 1 predicate resolution16 1 predicate resolution
16 1 predicate resolution
 
15 predicate
15 predicate15 predicate
15 predicate
 
09 heuristic search
09 heuristic search09 heuristic search
09 heuristic search
 
08 uninformed search
08 uninformed search08 uninformed search
08 uninformed search
 

Recently uploaded

What Causes 'Trans Failsafe Prog' to Trigger in BMW X5
What Causes 'Trans Failsafe Prog' to Trigger in BMW X5What Causes 'Trans Failsafe Prog' to Trigger in BMW X5
What Causes 'Trans Failsafe Prog' to Trigger in BMW X5
European Service Center
 
What Does the Active Steering Malfunction Warning Mean for Your BMW
What Does the Active Steering Malfunction Warning Mean for Your BMWWhat Does the Active Steering Malfunction Warning Mean for Your BMW
What Does the Active Steering Malfunction Warning Mean for Your BMW
Tanner Motors
 
What Exactly Is The Common Rail Direct Injection System & How Does It Work
What Exactly Is The Common Rail Direct Injection System & How Does It WorkWhat Exactly Is The Common Rail Direct Injection System & How Does It Work
What Exactly Is The Common Rail Direct Injection System & How Does It Work
Motor Cars International
 
Why Is Your BMW X3 Hood Not Responding To Release Commands
Why Is Your BMW X3 Hood Not Responding To Release CommandsWhy Is Your BMW X3 Hood Not Responding To Release Commands
Why Is Your BMW X3 Hood Not Responding To Release Commands
Dart Auto
 
一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理
一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理
一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理
psavhef
 
Ec460b lc Excavator Volvo Service Repair.pdf
Ec460b lc Excavator Volvo Service Repair.pdfEc460b lc Excavator Volvo Service Repair.pdf
Ec460b lc Excavator Volvo Service Repair.pdf
Excavator
 
Why Isn't Your BMW X5's Comfort Access Functioning Properly Find Out Here
Why Isn't Your BMW X5's Comfort Access Functioning Properly Find Out HereWhy Isn't Your BMW X5's Comfort Access Functioning Properly Find Out Here
Why Isn't Your BMW X5's Comfort Access Functioning Properly Find Out Here
Masters European & Gapanese Auto Repair
 
5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention
5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention
5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention
Bertini's German Motors
 
一比一原版BC毕业证波士顿学院毕业证成绩单如何办理
一比一原版BC毕业证波士顿学院毕业证成绩单如何办理一比一原版BC毕业证波士顿学院毕业证成绩单如何办理
一比一原版BC毕业证波士顿学院毕业证成绩单如何办理
amvovau
 
What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...
What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...
What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...
Autohaus Service and Sales
 
Statistics5,c.xz,c.;c.;d.c;d;ssssss.pptx
Statistics5,c.xz,c.;c.;d.c;d;ssssss.pptxStatistics5,c.xz,c.;c.;d.c;d;ssssss.pptx
Statistics5,c.xz,c.;c.;d.c;d;ssssss.pptx
coc7987515756
 
Antique Plastic Traders Company Profile
Antique Plastic Traders Company ProfileAntique Plastic Traders Company Profile
Antique Plastic Traders Company Profile
Antique Plastic Traders
 
Ec330B Lc Excavator Volvo Service Repair.pdf
Ec330B Lc Excavator Volvo Service Repair.pdfEc330B Lc Excavator Volvo Service Repair.pdf
Ec330B Lc Excavator Volvo Service Repair.pdf
Excavator
 
Empowering Limpopo Entrepreneurs Consulting SMEs.pptx
Empowering Limpopo Entrepreneurs  Consulting SMEs.pptxEmpowering Limpopo Entrepreneurs  Consulting SMEs.pptx
Empowering Limpopo Entrepreneurs Consulting SMEs.pptx
Precious Mvulane CA (SA),RA
 
Core technology of Hyundai Motor Group's EV platform 'E-GMP'
Core technology of Hyundai Motor Group's EV platform 'E-GMP'Core technology of Hyundai Motor Group's EV platform 'E-GMP'
Core technology of Hyundai Motor Group's EV platform 'E-GMP'
Hyundai Motor Group
 
gtyccccccccccccccccccccccccccccccccccccccccccccccccccccccc
gtycccccccccccccccccccccccccccccccccccccccccccccccccccccccgtyccccccccccccccccccccccccccccccccccccccccccccccccccccccc
gtyccccccccccccccccccccccccccccccccccccccccccccccccccccccc
4thzenzstar
 
Wondering if Your Mercedes EIS is at Fault Here’s How to Tell
Wondering if Your Mercedes EIS is at Fault Here’s How to TellWondering if Your Mercedes EIS is at Fault Here’s How to Tell
Wondering if Your Mercedes EIS is at Fault Here’s How to Tell
Vic Auto Collision & Repair
 
Things to remember while upgrading the brakes of your car
Things to remember while upgrading the brakes of your carThings to remember while upgrading the brakes of your car
Things to remember while upgrading the brakes of your car
jennifermiller8137
 
Tyre Industrymarket overview with examples of CEAT
Tyre Industrymarket overview with examples of CEATTyre Industrymarket overview with examples of CEAT
Tyre Industrymarket overview with examples of CEAT
kshamashah95
 

Recently uploaded (19)

What Causes 'Trans Failsafe Prog' to Trigger in BMW X5
What Causes 'Trans Failsafe Prog' to Trigger in BMW X5What Causes 'Trans Failsafe Prog' to Trigger in BMW X5
What Causes 'Trans Failsafe Prog' to Trigger in BMW X5
 
What Does the Active Steering Malfunction Warning Mean for Your BMW
What Does the Active Steering Malfunction Warning Mean for Your BMWWhat Does the Active Steering Malfunction Warning Mean for Your BMW
What Does the Active Steering Malfunction Warning Mean for Your BMW
 
What Exactly Is The Common Rail Direct Injection System & How Does It Work
What Exactly Is The Common Rail Direct Injection System & How Does It WorkWhat Exactly Is The Common Rail Direct Injection System & How Does It Work
What Exactly Is The Common Rail Direct Injection System & How Does It Work
 
Why Is Your BMW X3 Hood Not Responding To Release Commands
Why Is Your BMW X3 Hood Not Responding To Release CommandsWhy Is Your BMW X3 Hood Not Responding To Release Commands
Why Is Your BMW X3 Hood Not Responding To Release Commands
 
一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理
一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理
一比一原版SDSU毕业证圣地亚哥州立大学毕业证成绩单如何办理
 
Ec460b lc Excavator Volvo Service Repair.pdf
Ec460b lc Excavator Volvo Service Repair.pdfEc460b lc Excavator Volvo Service Repair.pdf
Ec460b lc Excavator Volvo Service Repair.pdf
 
Why Isn't Your BMW X5's Comfort Access Functioning Properly Find Out Here
Why Isn't Your BMW X5's Comfort Access Functioning Properly Find Out HereWhy Isn't Your BMW X5's Comfort Access Functioning Properly Find Out Here
Why Isn't Your BMW X5's Comfort Access Functioning Properly Find Out Here
 
5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention
5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention
5 Warning Signs Your BMW's Intelligent Battery Sensor Needs Attention
 
一比一原版BC毕业证波士顿学院毕业证成绩单如何办理
一比一原版BC毕业证波士顿学院毕业证成绩单如何办理一比一原版BC毕业证波士顿学院毕业证成绩单如何办理
一比一原版BC毕业证波士顿学院毕业证成绩单如何办理
 
What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...
What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...
What Does the PARKTRONIC Inoperative, See Owner's Manual Message Mean for You...
 
Statistics5,c.xz,c.;c.;d.c;d;ssssss.pptx
Statistics5,c.xz,c.;c.;d.c;d;ssssss.pptxStatistics5,c.xz,c.;c.;d.c;d;ssssss.pptx
Statistics5,c.xz,c.;c.;d.c;d;ssssss.pptx
 
Antique Plastic Traders Company Profile
Antique Plastic Traders Company ProfileAntique Plastic Traders Company Profile
Antique Plastic Traders Company Profile
 
Ec330B Lc Excavator Volvo Service Repair.pdf
Ec330B Lc Excavator Volvo Service Repair.pdfEc330B Lc Excavator Volvo Service Repair.pdf
Ec330B Lc Excavator Volvo Service Repair.pdf
 
Empowering Limpopo Entrepreneurs Consulting SMEs.pptx
Empowering Limpopo Entrepreneurs  Consulting SMEs.pptxEmpowering Limpopo Entrepreneurs  Consulting SMEs.pptx
Empowering Limpopo Entrepreneurs Consulting SMEs.pptx
 
Core technology of Hyundai Motor Group's EV platform 'E-GMP'
Core technology of Hyundai Motor Group's EV platform 'E-GMP'Core technology of Hyundai Motor Group's EV platform 'E-GMP'
Core technology of Hyundai Motor Group's EV platform 'E-GMP'
 
gtyccccccccccccccccccccccccccccccccccccccccccccccccccccccc
gtycccccccccccccccccccccccccccccccccccccccccccccccccccccccgtyccccccccccccccccccccccccccccccccccccccccccccccccccccccc
gtyccccccccccccccccccccccccccccccccccccccccccccccccccccccc
 
Wondering if Your Mercedes EIS is at Fault Here’s How to Tell
Wondering if Your Mercedes EIS is at Fault Here’s How to TellWondering if Your Mercedes EIS is at Fault Here’s How to Tell
Wondering if Your Mercedes EIS is at Fault Here’s How to Tell
 
Things to remember while upgrading the brakes of your car
Things to remember while upgrading the brakes of your carThings to remember while upgrading the brakes of your car
Things to remember while upgrading the brakes of your car
 
Tyre Industrymarket overview with examples of CEAT
Tyre Industrymarket overview with examples of CEATTyre Industrymarket overview with examples of CEAT
Tyre Industrymarket overview with examples of CEAT
 

Lecture7 xing fei-fei

  • 1. Eric Xing © Eric Xing @ CMU, 2006-2010 1 Machine Learning Data visualization and dimensionality reduction Eric Xing Lecture 7, August 13, 2010
  • 2. Eric Xing © Eric Xing @ CMU, 2006-2010 2 Text document retrieval/labelling  Represent each document by a high-dimensional vector in the space of words
  • 3. Eric Xing © Eric Xing @ CMU, 2006-2010 3 Image retrieval/labelling             = nx x x x  2 1
  • 4. Eric Xing © Eric Xing @ CMU, 2006-2010 4 Dimensionality Bottlenecks  Data dimension  Sensor response variables X:  1,000,000 samples of an EM/Acoustic field on each of N sensors  10242 pixels of a projected image on a IR camera sensor  N2 expansion factor to account for all pairwise correlations  Information dimension  Number of free parameters describing probability densities f(X) or f(S|X)  For known statistical model: info dim = model dim  For unknown model: info dim = dim of density approximation  Parametric-model driven dimension reduction  DR by sufficiency, DR by maximum likelihood  Data-driven dimension reduction  Manifold learning, structure discovery
  • 5. Eric Xing © Eric Xing @ CMU, 2006-2010 5 Intuition: how does your brain store these pictures?
  • 6. Eric Xing © Eric Xing @ CMU, 2006-2010 6 Brain Representation
  • 7. Eric Xing © Eric Xing @ CMU, 2006-2010 7 Brain Representation  Every pixel?  Or perceptually meaningful structure?  Up-down pose  Left-right pose  Lighting direction So, your brain successfully reduced the high-dimensional inputs to an intrinsically 3-dimensional manifold!
  • 8. Eric Xing © Eric Xing @ CMU, 2006-2010 8 Manifold Embedding (Metric) data geometry (Non-metric) information geometry Domain are i.i.d. samples from Two Geometries to Consider
  • 9. Eric Xing © Eric Xing @ CMU, 2006-2010 9  Data-driven projection to lower dimensional subsapce  Extract low-dim structure from high-dim data  Data may lie on curved (but locally linear) subspace [1] Josh .B. Tenenbaum, Vin de Silva, and John C. Langford “A Global Geometric Framework for Nonlinear Dimensionality Reduction” Science, 22 Dec 2000. [2] Jose Costa, Neal Patwari and Alfred O. Hero, “Distributed Weighted Multidimensional Scaling for Node Localization in Sensor Networks”, IEEE/ACM Trans. Sensor Networks, to appear 2005. [3] Misha Belkin and Partha Niyogi, “Laplacian eigenmaps for dimensionality reduction and data representation,” Neural Computation, 2003. Data-driven DR
  • 10. Eric Xing © Eric Xing @ CMU, 2006-2010 10 What is a Manifold?  A manifold is a topological space which is locally Euclidean.  Represents a very useful and challenging unsupervised learning problem.  In general, any object which is nearly "flat" on small scales is a manifold.
  • 11. Eric Xing © Eric Xing @ CMU, 2006-2010 11 Manifold Learning  Discover low dimensional structures (smooth manifold) for data in high dimension.  Linear Approaches  Principal component analysis.  Multi dimensional scaling.  Non Linear Approaches  Local Linear Embedding  ISOMAP  Laplacian Eigenmap.
  • 12. Eric Xing © Eric Xing @ CMU, 2006-2010 12 Principal component analysis  Areas of variance in data are where items can be best discriminated and key underlying phenomena observed  If two items or dimensions are highly correlated or dependent  They are likely to represent highly related phenomena  We want to combine related variables, and focus on uncorrelated or independent ones, especially those along which the observations have high variance  We look for the phenomena underlying the observed covariance/co- dependence in a set of variables  These phenomena are called “factors” or “principal components” or “independent components,” depending on the methods used  Factor analysis: based on variance/covariance/correlation  Independent Component Analysis: based on independence
  • 13. Eric Xing © Eric Xing @ CMU, 2006-2010 13 An example:
  • 14. Eric Xing © Eric Xing @ CMU, 2006-2010 14 Principal Component Analysis  The new variables/dimensions  Are linear combinations of the original ones  Are uncorrelated with one another  Orthogonal in original dimension space  Capture as much of the original variance in the data as possible  Are called Principal Components  Orthogonal directions of greatest variance in data  Projections along PC1 discriminate the data most along any one axis  First principal component is the direction of greatest variability (covariance) in the data  Second is the next orthogonal (uncorrelated) direction of greatest variability  So first remove all the variability along the first component, and then find the next direction of greatest variability  And so on … Original Variable A OriginalVariableB PC 1 PC 2
  • 15. Eric Xing © Eric Xing @ CMU, 2006-2010 15 Computing the Components  Projection of vector x onto an axis (dimension) u is uTx  Direction of greatest variability is that in which the average square of the projection is greatest: Maximize uTXXTu s.t uTu = 1 Construct Langrangian uTXXTu – λuTu Vector of partial derivatives set to zero xxTu – λu = (xxT – λI) u = 0 As u ≠ 0 then u must be an eigenvector of XXT with eigenvalue λ  λ is the principal eigenvalue of the correlation matrix C= XXT  The eigenvalue denotes the amount of variability captured along that dimension
  • 16. Eric Xing © Eric Xing @ CMU, 2006-2010 16 Computing the Components  Similarly for the next axis, etc.  So, the new axes are the eigenvectors of the matrix of correlations of the original variables, which captures the similarities of the original variables based on how data samples project to them  Geometrically: centering followed by rotation  Linear transformation
  • 17. Eric Xing © Eric Xing @ CMU, 2006-2010 17  For symmetric matrices, eigenvectors for distinct eigenvalues are orthogonal  All eigenvalues of a real symmetric matrix are real.  All eigenvalues of a positive semidefinite matrix are non- negative ℜ∈⇒==− λλ T SSand0if IS 0vSvifthen,0, ≥⇒=≥ℜ∈∀ λλSwww Tn 02121212121 =•⇒≠= vvvSv λλλ and,},{},{},{ Eigenvalues & Eigenvectors
  • 18. Eric Xing © Eric Xing @ CMU, 2006-2010 18  Let be a square matrix with m linearly independent eigenvectors (a “non-defective” matrix)  Theorem: Exists an eigen decomposition (cf. matrix diagonalization theorem)  Columns of U are eigenvectors of S  Diagonal elements of are eigenvalues of Eigen/diagonal Decomposition diagonal Unique for distinct eigen- values
  • 19. Eric Xing © Eric Xing @ CMU, 2006-2010 19 PCs, Variance and Least-Squares  The first PC retains the greatest amount of variation in the sample  The kth PC retains the kth greatest fraction of the variation in the sample  The kth largest eigenvalue of the correlation matrix C is the variance in the sample along the kth PC  The least-squares view: PCs are a series of linear least squares fits to a sample, each orthogonal to all previous ones
  • 20. Eric Xing © Eric Xing @ CMU, 2006-2010 20 0 5 10 15 20 25 PC1 PC2 PC3 PC4 PC5 PC6 PC7 PC8 PC9 PC10 Variance(%) How Many PCs?  For n original dimensions, sample covariance matrix is nxn, and has up to n eigenvectors. So n PCs.  Where does dimensionality reduction come from? Can ignore the components of lesser significance. You do lose some information, but if the eigenvalues are small, you don’t lose much  n dimensions in original data  calculate n eigenvectors and eigenvalues  choose only the first p eigenvectors, based on their eigenvalues  final data set has only p dimensions
  • 21. Eric Xing © Eric Xing @ CMU, 2006-2010 21 Application: querying text doc.
  • 22. Eric Xing © Eric Xing @ CMU, 2006-2010 22 Within .40 threshold K is the number of singular values used
  • 23. Eric Xing © Eric Xing @ CMU, 2006-2010 23 Summary:  Principle  Linear projection method to reduce the number of parameters  Transfer a set of correlated variables into a new set of uncorrelated variables  Map the data into a space of lower dimensionality  Form of unsupervised learning  Properties  It can be viewed as a rotation of the existing axes to new positions in the space defined by original variables  New axes are orthogonal and represent the directions with maximum variability  Application: In many settings in pattern recognition and retrieval, we have a feature-object matrix.  For text, the terms are features and the docs are objects.  Could be opinions and users …  This matrix may be redundant in dimensionality.  Can work with low-rank approximation.  If entries are missing (e.g., users’ opinions), can recover if dimensionality is low.
  • 24. Eric Xing © Eric Xing @ CMU, 2006-2010 24 Going beyond  What is the essence of the C matrix?  The elements in C captures some kind of affinity between a pair of data points in the semantic space  We can replace it with any reasonable affinity measure  E.g., MDS  E.g., the geodistance ISOMAP ( ) matrixdistance: 2 ijji xxD −=
  • 25. Eric Xing © Eric Xing @ CMU, 2006-2010 25  Constructing neighbourhood graph G  For each pair of points in G, Computing shortest path distances ---- geodesic distances.  Use Dijkstra's or Floyd's algorithm  Apply kernel PCA for C given by the centred matrix of squared geodesic distances.  Project test points onto principal components as in kernel PCA. Nonlinear DR – Isomap [Josh. Tenenbaum, Vin de Silva, John langford 2000]
  • 26. Eric Xing © Eric Xing @ CMU, 2006-2010 3D data 2D coord chart Error vs. dimensionality of coordinate chart “Swiss Roll” dataset
  • 27. Eric Xing © Eric Xing @ CMU, 2006-2010 27 PCA, MD vs ISOMAP  The residual variance of PCA (open triangles), MDS (open circles), and Isomap
  • 28. Eric Xing © Eric Xing @ CMU, 2006-2010 28 ISOMAP algorithm Pros/Cons Advantages:  Nonlinear  Globally optimal  Guarantee asymptotically to recover the true dimensionality Drawback:  May not be stable, dependent on topology of data  As N increases, pair wise distances provide better approximations to geodesics, but cost more computation
  • 29. Eric Xing © Eric Xing @ CMU, 2006-2010 29 Local Linear Embedding (a.k.a LLE)  LLE is based on simple geometric intuitions.  Suppose the data consist of N real-valued vectors Xi, each of dimensionality D.  Each data point and its neighbors expected to lie on or close to a locally linear patch of the manifold.
  • 30. Eric Xing © Eric Xing @ CMU, 2006-2010 30 Steps in LLE algorithm  Assign neighbors to each data point  Compute the weights Wij that best linearly reconstruct the data point from its neighbors, solving the constrained least-squares problem.  Compute the low-dimensional embedding vectors best reconstructed by Wij. iX  iY 
  • 31. Eric Xing © Eric Xing @ CMU, 2006-2010 31 Fit locally, Think Globally From Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul ∑ ∑−=Φ i j jijYWYY 2 ||)( 
  • 32. Eric Xing © Eric Xing @ CMU, 2006-2010 32 Super-Resolution Through Neighbor Embedding [Yeung et al CVPR 2004] Training Xsi Training Ysi Testing Xt Testing Yt ?
  • 33. Eric Xing © Eric Xing @ CMU, 2006-2010 33 Training Xsi Training Ysi High dimensional Manifold Low dimensional Manifold Intuition  Patches of the image lie on a manifold
  • 34. Eric Xing © Eric Xing @ CMU, 2006-2010 34 Algorithm 1. Get feature vectors for each low resolution training patch. 2. For each test patch feature vector find K nearest neighboring feature vectors of training patches. 3. Find optimum weights to express each test patch vector as a weighted sum of its K nearest neighbor vectors. 4. Use these weights for reconstruction of that test patch in high resolution.
  • 35. Eric Xing © Eric Xing @ CMU, 2006-2010 35 Training Xsi Training Ysi Testing Xt Testing Yt Results
  • 36. Eric Xing © Eric Xing @ CMU, 2006-2010 36 Summary:  Principle  Linear and nonlinear projection method to reduce the number of parameters  Transfer a set of correlated variables into a new set of uncorrelated variables  Map the data into a space of lower dimensionality  Form of unsupervised learning  Applications  PCA and Latent semantic indexing for text mining  Isomap and Nonparametric Models of Image Deformation  LLE and Isomap Analysis of Spectra and Colour Images  Image Spaces and Video Trajectories: Using Isomap to Explore Video Sequences  Mining the structural knowledge of high-dimensional medical data using isomap Isomap Webpage: http://isomap.stanford.edu/
  • 37. Applying PCA and LDA: Eigen-faces and Fisher-faces L. Fei-Fei Computer Science Dept. Stanford University
  • 38. Machine learning in computer vision • Aug 13, Lecture 7: Dimensionality reduction, Manifold learning – Eigen- and Fisher- faces – Applications to object representation 8/8/2010 38L. Fei-Fei, Dragon Star 2010, Stanford
  • 39. The Space of Faces • An image is a point in a high dimensional space – An N x M image is a point in RNM – We can define vectors in this space as we did in the 2D case += [Thanks to Chuck Dyer, Steve Seitz, Nishino]
  • 40. Key Idea }ˆ{ P RLx=χ• Images in the possible set are highly correlated. • So, compress them to a low-dimensional subspace that captures key appearance characteristics of the visual DOFs. • EIGENFACES: [Turk and Pentland] USE PCA!
  • 41. Principal Component Analysis (PCA) • PCA is used to determine the most representing features among data points. – It computes the p-dimensional subspace such that the projection of the data points onto the subspace has the largest variance among all p-dimensional subspaces.
  • 42. Illustration of PCA One projection PCA projection x1 x2 1 2 3 4 5 6 x1 x2 1 2 3 4 5 6 X1’
  • 43. Illustration of PCA x1 x2 1st principal component 2rd principal component
  • 44. Mathematical Formulation Find a transformation, W, m-dimensional n-dimensionalOrthonormal Total scatter matrix: Wopt corresponds to m eigen- vectors of ST
  • 45. Eigenfaces • PCA extracts the eigenvectors of A – Gives a set of vectors v1, v2, v3, ... – Each one of these vectors is a direction in face space • what do these look like?
  • 46. Projecting onto the Eigenfaces • The eigenfaces v1, ..., vK span the space of faces – A face is converted to eigenface coordinates by
  • 47. Algorithm 1. Align training images x1, x2, …, xN 2. Compute average face u = 1/N Σ xi 3. Compute the difference image φi = xi – u Training Note that each image is formulated into a long vector!
  • 48. Algorithm Testing 1. Projection in Eigenface Projection ωi = W(X – u), W = {eigenfaces} 2. Compare projections ST = 1/NΣ φi φi T = BBT, B=[φ1, φ2 … φN] 4. Compute the covariance matrix (total scatter matrix) 5. Compute the eigenvectors of the covariance matrix , W
  • 49. Illustration of Eigenfaces These are the first 4 eigenvectors from a training set of 400 images (ORL Face Database). They look like faces, hence called Eigenface. The visualization of eigenvectors:
  • 50. Eigenfaces look somewhat like generic faces.
  • 52. Only selecting the top P eigenfaces  reduces the dimensionality. Fewer eigenfaces result in more information loss, and hence less discrimination between faces. Reconstruction and Errors P = 4 P = 200 P = 400
  • 53. Summary for PCA and Eigenface • Non-iterative, globally optimal solution • PCA projection is optimal for reconstruction from a low dimensional basis, but may NOT be optimal for discrimination…
  • 54. Linear Discriminant Analysis (LDA) • Using Linear Discriminant Analysis (LDA) or Fisher’s Linear Discriminant (FLD) • Eigenfaces attempt to maximise the scatter of the training images in face space, while Fisherfaces attempt to maximise the between class scatter, while minimising the within class scatter.
  • 55. Illustration of the Projection Poor Projection Good Projection x1 x2 x1 x2  Using two classes as example:
  • 57. Variables • N Sample images: • c classes: • Average of each class: • Total average: { }Nxx ,,1  { }cχχ ,,1  ∑= ∈ ikx k i i x N χ µ 1 ∑= = N k kx N 1 1 µ
  • 58. Scatters • Scatter of class i: ( )( )T ik x iki xxS ik µµ χ −∑ −= ∈ ∑= = c i iW SS 1 ( )( )∑ −−= = c i T iiiBS 1 µµµµχ BWT SSS += • Within class scatter: • Between class scatter: • Total scatter:
  • 60. Mathematical Formulation (1) After projection: Between class scatter (of y’s): Within class scatter (of y’s): k T k xWy = WSWS B T B = ~ WSWS W T W = ~
  • 61. Mathematical Formulation (2) • The desired projection: WSW WSW S S W W T B T W B opt WW maxarg~ ~ maxarg == miwSwS iWiiB ,,1 == λ • How is it found ?  Generalized Eigenvectors Data dimension is much larger than the number of samples The matrix is singular: Nn >> ( ) cNSRank W −≤WS
  • 62. Fisherface (PCA+FLD) • Project with FLD to space • Project with PCA to space 1−c k T fldk zWy = WWSWW WWSWW W pcaW T pca T pcaB T pca T fld W maxarg= cN − k T pcak xWz = WSWW T T pca W maxarg=
  • 64. Results: Eigenface vs. Fisherface (1) • Variation in Facial Expression, Eyewear, and Lighting • Input: 160 images of 16 people • Train: 159 images • Test: 1 image With glasses Without glasses 3 Lighting conditions 5 expressions
  • 66. discussion • Removing the first three principal components results in better performance under variable lighting conditions • The Firsherface methods had error rates lower than the Eigenface method for the small datasets tested.
  • 67. Manifold Learning for Object Representation L. Fei-Fei Computer Science Dept. Stanford University
  • 68. Machine learning in computer vision • Aug 13, Lecture 7: Dimensionality reduction, Manifold learning – Eigen- and Fisher- faces – Applications to object representation (slides courtesy to David Thompson) 8/8/2010 68L. Fei-Fei, Dragon Star 2010, Stanford
  • 70. appearance variation manifolds in vision images from hormel corp.
  • 71. deformation manifolds in vision images from www.golfswingphotos.com
  • 72. Find a low-D basis for describing high-D data. X ~= X' S.T. dim(X') << dim(X) uncovers the intrinsic dimensionality manifold learning
  • 73. If we knew all pairwise distances… Chicago Raleigh Boston Seattle S.F. Austin Orlando Chicago 0 Raleigh 641 0 Boston 851 608 0 Seattle 1733 2363 2488 0 S.F. 1855 2406 2696 684 0 Austin 972 1167 1691 1764 1495 0 Orlando 994 520 1105 2565 2458 1015 0 Distances calculated with geobytes.com/CityDistanceTool
  • 74. Multidimensional Scaling (MDS) For n data points, and a distance matrix D, Dij = ...we can construct a m-dimensional space to preserve inter-point distances by using the top eigenvectors of D scaled by their eigenvalues j i
  • 76. Actual plot of cities
  • 79. 1. data compression 2. “curse of dimensionality” 3. de-noising 4. visualization 5. reasonable distance metrics why do manifold learning?
  • 83. Isomap for images  Build a data graph G.  Vertices: images  (u,v) is an edge iff SSD(u,v) is small  For any two images, we approximate the distance between them with the “shortest path” on G
  • 84. Isomap 1. Build a sparse graph with K-nearest neighbors Dg = (distance matrix is sparse)
  • 85. Isomap 2. Infer other interpoint distances by finding shortest paths on the graph (Dijkstra's algorithm). Dg =
  • 86. Isomap shortest-distance on a graph is easy to compute
  • 88. - preserves global structure - few free parameters - sensitive to noise, noise edges - computationally expensive (dense matrix eigen-reduction) Isomap: pro and con
  • 90. Find a mapping to preserve local linear relationships between neighbors Locally Linear Embedding
  • 92. 1. Find weight matrix W of linear coefficients: Enforce sum-to-one constraint. LLE: Two key steps
  • 93. 2. Find projected vectors Y to minimize reconstruction error must solve for whole dataset simultaneously LLE: Two key steps
  • 98. - no local minima, one free parameter - incremental & fast - simple linear algebra operations - can distort global structure LLE: pro and con