SlideShare a Scribd company logo
1 of 122
Download to read offline
Some important information on the course
Lecturer: Jorge M. Mendes (jmm@novaims.unl.pt)
Lectures timetable: Tuesdays and Thursdays, 3:30pm to 4:15pm
Assessment rules: check syllabus
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 1 / 122
Some important information on the course
Textbooks
Everitt, B. and Hothorn, T. (2011). An Introduction to Applied Multivariate
Analysis with R, Springer.
Johnson, R.A and Wichern (2007). D. W., Applied Multivariate Statistical
Analysis, 6th edition, Pearson Prentice Hall.
online tools and references: https://www.datacamp.com
online tools and references: http://www.statmethods.net
And...
Reis, E. (1997), Estat´ıstica Multivariada Aplicada, Edi¸c˜oes S´ılabo
Sharma, S. (1996). Applied Multivariate Techniques, John Wiley & Sons
Timm, N. H. (2002). Applied Multivariate Analysis, Springer
Other things: lecture slides, data files, etc.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 2 / 122
Some important information on the course
Course contents
Data analysis basics: organisation of multivariate data; parameters and
descriptive statistics; matrix manipulations for sample statistics; review of
matrix definitions and operations
R tutorial; fundamentals on data manipulation on R
Graphical display of multivariate data; Linear combinations of random variables
The multivariate normal distribution
Principal components analysis
Factor analysis
Discriminant analysis
Cluster analysis
Canonical correlation analysis
Multidimensional scaling
... eventually something else!
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 3 / 122
Lecture 1
Multivariate Data Analysis Basics
Jorge M. Mendes
Nova Information Management School
Universidade Nova de Lisboa
2015
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 4 / 122
First things first
This first lesson will introduce the basic tools to deal with multivariate data
and multivariate techniques.
Topics covered include multivariate sample statistics and spectral
decomposition of a matrix.
After successfully completing this lesson, you should be able to:
Understand the notation used in multivariate analysis;
Understand the organization of multivariate data using vector and matrix
notation;
Perform basic operations on matrices and vectors;
Interpret measures of central tendency, dispersion, and association;
Calculate sample means, variances, covariances, and correlations using a hand
calculator;
Compute the spectral decomposition of a matrix (eigenvalues, eigenvectors,
etc.)
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 5 / 122
Multivariate data analysis basics Organisation of multivariate data
Subsection 1
Organisation of multivariate data
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 6 / 122
Multivariate data analysis basics Organisation of multivariate data
Introduction
Univariate statistics is concerned with random scalar variable Y .
In multivariate analysis, we are concerned with the joint analysis of
multiple dependent variables.
These variables can be represented using matrices and vectors.
This provides simplification of notation and a format for expressing
important formulas.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 7 / 122
Multivariate data analysis basics Organisation of multivariate data
Example
Suppose that we measure the variables x1 = height (cm), x2 = left forearm
length (cm) and x3 = left foot length for participants in a study of the
physical characteristics of adult humans.
These three variables can be represented in the following column vector:
x =


x1
x2
x3


Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 8 / 122
Multivariate data analysis basics Organisation of multivariate data
Example
The observed data for a specific individual, say the i-th individual, might also
be represented in an analogous vector.
Suppose that the i-th person in the sample has height = 175 cm, forearm
length = 25.5 cm and foot length = 27 cm.
In vector notation these observed data could be written as:
xi =


xi1
xi2
xi3

 =


175
25.5
27.0


Notice the use and placement of the subscript i to represent the i-th
individual.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 9 / 122
Multivariate data analysis basics Organisation of multivariate data
Definitions of matrix and vector
A matrix is two-dimensional array of numbers of formulas.
A vector is a matrix with either only one column or only one row. A column
vector has only one column. A row vector has only one row.
The dimension of a matrix is expressed as number of rows × number of
columns.
For instance, a matrix with 10 rows and 3 columns is said to be a 10 × 3
matrix. The vectors written in previous Example are 3 × 1 matrices.
A square matrix is one for which the numbers of rows and columns are the
same. For instance, a 4 × 4 matrix is a square matrix.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 10 / 122
Multivariate data analysis basics Organisation of multivariate data
The data matrix in multivariate problems
Usually the observed data are represented by a matrix in which the rows are
observations and the columns are variables.
This is exactly the way the data are normally prepared for statistical software.
The usual notation is
n: the number of observed units (people, animals, companies, etc.)
p: number of variables measured on each unit.
Thus the data matrix will be an n × p matrix.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 11 / 122
Multivariate data analysis basics Organisation of multivariate data
Example
Suppose that we have scores for n = 6 college students who have taken the
verbal and the science subtests of a College Qualification test (CQT).
We have p = 2 variables: (1) the verbal score and (2) the science score for
each student.
The data matrix is the following 6 × 2 matrix:
X =








41 26
39 26
53 21
67 33
61 27
67 29








Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 12 / 122
Multivariate data analysis basics Organisation of multivariate data
Example
In the matrix just given, the first column gives the data for x1 = verbal score
whereas the second column gives data for x2 = science score.
Each row gives data for a student in the sample.
To repeat - the rows are observations, the columns are variables.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 13 / 122
Multivariate data analysis basics Organisation of multivariate data
Notation notes
Note that we have used a small x to denote the vector of variables in first
example and a large X to represent the data matrix in second example.
It should also be noted that, in matrix terms, the i-th row in the data matrix
X is the transpose of the data vector
xi =
xi1
xi2
, as we defined data vectors in first Example.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 14 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Subsection 2
Parameters and descriptive statistics
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 15 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Introduction
Three aspects of the data are of importance, the first two of which you should
already be familiar with from univariate statistics. These are:
Central Tendency. What is a typical value for each variable?
Dispersion. How far apart are the individual observations from a central value
for a given variable?
Association. This might (or might not!) be a new measure for you. When
more than one variable are studied together how does each variable relate to
the remaining variables? How are the variables simultaneously related to one
another? Are they positively or negatively related?
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 16 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Population parameters and sample statistics
Statistics, as a subject matter, is the science and art of using sample
information to make generalizations about populations.
A population is the collection of all people, plants, animals, or objects of
interest about which we wish to make statistical inferences (generalizations).
A population parameter is a numerical characteristic of a population. In
nearly all statistical problems we do not know the value of a parameter
because we do not measure the entire population. We use sample data to
make an inference about the value of a parameter.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 17 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Population parameters and sample statistics
A sample is the subset of the population that we actually measure or observe.
A sample statistic is a numerical characteristic of a sample. A sample
statistic estimates the unknown value of a population parameter.
Information collected from sample statistic is sometimes referred to as
descriptive statistic.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 18 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Notation
Here are the notation that will be used:
Xij : Observation for variable j in subject i.
p: Number of variables
n: Number of subjects
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 19 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example: USDA Women’s Health Survey
Let us take a look at an example.
In 1985, the USDA commissioned a study of women’s nutrition.
Nutrient intake was measured for a random sample of 737 women aged 25-50
years.
The following variables were measured:
Calcium (mg)
Iron (mg)
Protein (g)
Vitamin A (µg)
Vitamin C (mg)
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 20 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example: USDA Women’s Health Survey
As a notation example let’s use the women’s nutrition data.
p = 5 nutritional outcomes
n = 737 subjects
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 21 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example: USDA Women’s Health Survey
In multivariate statistics we will always be working with vectors of
observations.
So in this case we are going to arrange the data for the p variables on each
subject into a vector.
In the expression below, Xi is the vector of observations for the i-th subject,
i = 1, ..., n(= 737).
Therefore, the data for the j-th variable will be located in the j-th element of
this subject’s vector, j = 1, ..., p(= 5).
Xi =





Xi1
Xi2
...
Xip





Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 22 / 122
Multivariate data analysis basics Parameters and descriptive statistics
The mean
Throughout this course, we’ll use the ordinary notations for the mean of a
variable.
That is, the symbol µ is used to represent a (theoretical) population mean
and the symbol ¯x is used to represent a sample mean computed from
observed data.
In the multivariate setting, we add subscripts to these symbols to indicate the
specific variable for which the mean is being given.
For instance, µ1 represents the population mean for variable x1 and ¯x1
denotes a sample mean based on observed data for variable x1.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 23 / 122
Multivariate data analysis basics Parameters and descriptive statistics
The mean
The population mean is the measure of central tendency for the population.
Here, the population mean for variable j is
µj = E(Xij )
The notation E stands for statistical expectation; here E(Xij ) is the mean of
Xij over all members of the population, or equivalently, over all random draws
from a stochastic model.
For example, µj = E(Xij ) may be the mean of a normal variable.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 24 / 122
Multivariate data analysis basics Parameters and descriptive statistics
The mean
The population mean µj for variable j can be estimated by the sample mean
¯xj =
1
n
n
i=1
Xij
Note: the sample mean ¯xj , because it is a function of our random data is
also going to have a mean itself.
In fact, the population mean of the sample mean is equal to population mean
µj ; i.e.,
E(¯xj ) = µj
Therefore, the ¯xj is unbiased for µj .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 25 / 122
Multivariate data analysis basics Parameters and descriptive statistics
The mean
Another way of saying this is that the mean of the ¯xj ’s over all possible
samples of size n is equal to µj .
Recall that the population mean vector is µ which is a collection of the
means for each of the population means for each of the different variables.
µ =





µ1
µ2
...
µp





Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 26 / 122
Multivariate data analysis basics Parameters and descriptive statistics
The mean
We can estimate this population mean vector, µ, by ¯x. This is obtained by
collecting the sample means from each of the variables in a single vector.
This is shown below.
¯x =





¯x1
¯x2
...
¯xp





=





1
n
n
i=1 Xi1
1
n
n
i=1 Xi2
...
1
n
n
i=1 Xip





=
1
n
n
i=1
Xi
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 27 / 122
Multivariate data analysis basics Parameters and descriptive statistics
The mean
Just as the sample means, ¯xj , for the individual variables are unbiased for
their respective population means, note that the sample mean vector is
unbiased for the population mean vector.
E(¯x) = E





¯x1
¯x2
...
¯xp





=





E(¯x1)
E(¯x2)
...
E(¯xp)





=





µ1
µ2
...
µp





= µ
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 28 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Dispersion: variance and standard deviation
A variance measures the degree of spread (dispersion) in a variable’s values.
Theoretically, a population variance is the average squared difference between
a variable’s values and the mean for that variable. The population variance
for variable xj is
σ2
j = E(xj − µj )2
Note that the squared residual (xj − µj )2
is a function of the random variable
Xij .
Therefore, the squared residual itself is random, and has a population mean.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 29 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Dispersion: variance and standard deviation
The population variance is thus the population mean of the squared residual.
We see that if the data tend to be far away from the mean, the squared
residual will tend to be large, and hence the population variance will also be
large.
Conversely, if the the data tend to be close to the mean, the squared residual
will tend to be small, and hence the population variance will also be small.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 30 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Dispersion: variance and standard deviation
The population variance σ2
j can be estimated by the sample variance
s2
j =
1
n − 1
n
i=1
(Xij − ¯xj )2
=
1
n − 1
n
i=1
X2
ij −
n
n − 1
¯x2
j
The first expression in this formula is most suitable for interpreting the
sample variance.
We see that it is a function of the squared residuals; that is, take difference
between the individual observations and their sample mean, and then square
the result.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 31 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Dispersion: variance and standard deviation
Here, we may observe that if tend to be far away from their sample means,
then the squared residuals and hence the sample variance will also tend to be
large.
If on the other hand, if the observations tend to be close to their respective
sample means, then the squared differences between the data and their means
will be small, resulting is a small sample variance value for that variable.
The last part of the expression above, gives the formula that is most
suitable for computation, either by hand or by a computer!
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 32 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Dispersion: variance and standard deviation
Since the sample variance is a function of the random data, the sample
variance itself is a random quantity, and so has a population mean.
In fact, the population mean of the sample variance is equal to the
population variance:
E(s2
j ) = σ2
j
That is, the sample variance s2
j is unbiased for the population variance σ2
j .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 33 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Dispersion: variance and standard deviation
Some textbooks use a sample variance formula derived using maximum
likelihood estimation principles. In this formula, the division is by n rather
than n − 1.
s2
j =
n
i=1(xij − ¯xj )2
n
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 34 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Suppose that we have observed the following n = 5 resting pulse rates: 64,
68, 74, 76, 78. The sample mean is ¯x = 64+68+74+76+78
5 = 72.
The maximum likelihood estimate of the variance is
s2
=
(64 − 72)2
+ (68 − 72)2
+ (74 − 72)2
+
5
+(76 − 72)2
+ (78 − 72)2
5
=
136
5
= 27.2
The standard deviation based in this method is s =
√
27.2 = 5.215.
The more commonly used variance estimate, the one given by statistical
software, would be 136
5−1 = 34. The standard deviation would be
s =
√
34 = 5.83.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 35 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Association is concerned with how each variable is related to the other
variable(s). In this case the first measure that we will consider is the
covariance between two variables j and k.
The population covariance is a measure of the association between pairs of
variables in a population. Here, the population covariance between variables j
and k is
σjk = E{(Xij − µj )(Xik − µk )}
Note that the product of the residuals (Xij − µj ) and (Xik − µk ) for variables
j and k, respectively, is a function of the random variables Xij and Xik .
Therefore, (Xij − µj )(Xik − µk ) is itself random, and has a population mean.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 36 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
The population covariance is defined to be the population mean of this
product of residuals. We see that if either both variables are greater than
their respective means, or if they are both less than their respective means,
then the product of the residuals will be positive.
Thus, if the value of variable j tends to be greater than its mean when the
value of variable k is larger than its mean, and if the value of variable j tends
to be less than its mean when the value of variable k is smaller than its
mean, then the covariance will be positive.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 37 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Positive population covariances mean that the two variables are positively
associated; variable j tends to increase with increasing values of variable k
and vice-versa.
Negative association can also occur. If one variable tends to be greater than
its mean when the other variable is less than its mean, the product of the
residuals will be negative, and you will obtain a negative population
covariance. Variable j will tend to decrease with increasing values of variable
k.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 38 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
The population covariance σjk between variables j and k can be estimated by
the sample covariance. This can be calculated using the formula below:
sjk =
1
n − 1
n
i=1
(Xij − ¯xj )(Xik − ¯xk ) =
1
n − 1
n
i=1
Xij Xik −
n
n − 1
¯x2
j ¯x2
k
Just like in the formula for variance we have two expressions that make up
this formula. The first half of the formula is most suitable for understanding
the interpretation of the sample covariance, and the second half of the
formula is what is used for calculation.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 39 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Looking at the first half of the expression what you see inside the sum is the
product of the residual differences between variable j and its mean times the
residual differences between variable k and its mean.
We can see that if either both variables tend to be greater than their
respective means or less than their respective means, then the product of the
residuals will tend to be positive leading to a positive sample covariance.
Conversely if one variable takes values that are greater than its mean when
the opposite variable takes a value less than its mean, then the product will
take a negative value. In the end, when you add up this product over all of
the observations, you will end up with a negative covariance.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 40 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
So, in effect, a positive covariance would indicate a positive association
between the variables j and k. And a negative association is when the
covariance is negative.
For computational purposes we will use the second half of the formula. For
each subject, the product of the two variables is obtained, and then the
products are summed to obtain the first term in the numerator.
The second term in the numerator is obtained by taking the product of the
sums of variable over the n subjects, then dividing the results by the sample
size n. The difference between the first and second terms is then divided by
n − 1 to obtain the covariance value.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 41 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Again, sample covariance is a function of the random data, and hence, is
random itself.
As before, the population mean of the sample covariance sjk is equal the
population covariance σjk ; i.e.,
E(sjk ) = σjk
That is, the sample covariance sjk is unbiased for the population covariance
σjk .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 42 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
The sample covariance is a measure of the association between a pair of
variables:
sjk = 0 implies that the two variables are uncorrelated. (Note that this does
not necessarily imply independence, we’ll get back to this later.)
sjk > 0 implies that the two variables are positively correlated; i.e., values of
variable j tend to increase with increasing values of variable k. The larger the
covariance, the stronger the positive association between the two variables.
sjk < 0 implies that the two variables are negatively correlated; i.e., values of
variable j tend to decrease with increasing values of variable k. The smaller the
covariance, the stronger the negative association between the two variables.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 43 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Recall, that we had collected all of the population means of the p variables
into a mean vector.
Likewise, the population variances and covariances can be collected into the
population variance-covariance matrix. This is also known by the name of
population dispersion matrix.
Σ =





σ2
1 σ12 . . . σ1p
σ21 σ2
2 . . . σ2p
...
...
...
...
σp1 σp2 . . . σ2
p





Note that the population variances appear along the diagonal of this matrix,
and the covariance appear in the off-diagonal elements. So, the covariance
between variables j and k will appear in row j and column k of this matrix.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 44 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
The population variance-covariance matrix may be estimated by the sample
variance-covariance matrix.
The population variances and covariances in the above population
variance-covariance matrix are replaced by the corresponding sample
variances and covariances to obtain the sample variance-covariance matrix:
S =





s2
1 s12 . . . s1p
s21 s2
2 . . . s2p
...
...
...
...
sp1 sp2 . . . s2
p





Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 45 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Note that the sample variances appear along diagonal of this matrix and the
covariances appear in the off-diagonal elements. So the covariance between
variables j and k will appear in the jk-th element of this matrix.
Notes:
S (the sample variance-covariance matrix) is symmetric; i.e., sjk = skj .
S is unbiased for the population variance covariance matrix Σ ; i.e.,
E(S) =





E(s2
1 ) E(s12) . . . E(s1p)
E(s21) E(s2
2 ) . . . E(s2p)
...
...
...
...
E(sp1) E(sp2) . . . E(s2
p )





= Σ
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 46 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Since this matrix is a function of our random data, this means that the
elements of this matrix are also going to be random, and the matrix on the
whole is random as well.
The statement “σ2
is unbiased”means that the mean of each element of that
matrix is equal to the corresponding elements of the population.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 47 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
In matrix notation sample variance-covariance matrix may be computed used
the following expressions:
S =
1
n − 1
n
i=1
(Xi − ¯x)(Xi − ¯x) =
1
n − 1
n
i=1
Xi Xi −
n
n − 1
¯x¯x
Just as we have seen in the previous formulas, the first half of the formula is
used in interpretation, and the second half of the formula is what is used for
calculation purposes.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 48 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Measures of association: covariance and correlation
Looking at the second term you can see that the first term in the numerator
involves taking the data vector for each subject and multiplying by its
transpose. The resulting matrices are then added over the n subjects.
To obtain the second term in numerator, first compute the sum of the data
vectors over the n subjects, then take the resulting vector and multiply by its
transpose; then divide the resulting matrix by the squared number of subjects
n2
. Take the difference between the two terms in the numerator and divide
by n − 1.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 49 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Suppose that we have observed the following n = 4 observations for variables
x1 and x2.
x1 x2
6 3
10 4
12 7
12 6
The sample means are x1 = 10 and x2 = 5.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 50 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
The maximum likelihood estimate of the covariance is the average product of
deviations from the mean:
s12 =
(6 − 10)(3 − 5) + (10 − 10)(4 − 5)+
4
+(12 − 10)(7 − 5) + (12 − 10)(6 − 5)
4
=
8 + 0 + 4 + 2
4
= 3.5
The positive value reflects the fact that as x1 increases, x2 also tends to
increase.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 51 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Problem
Problem: The magnitude of the covariance value is not particularly helpful
as it is a function of the magnitudes (scales) of the two variables.
It is hard to distinguish the effects of the association between the two
variables from the effects of their dispersions.
Note, however, that the covariance between variables j and k must lie
between the product of the two component standard deviations of variables j
and k, and negative of that same product:
−sj sk ≤ sjk ≤ sj sk .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 52 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
In an undergraduate statistics class, n = 30 females reported their heights
(inches), and also measured their left forearm length (cm), left foot length
(cm), and head circumference (cm).
The sample variance-covariance matrix is the following:
Notice that the matrix has four rows and four columns because there are four
variables being considered. Also notice that the matrix is symmetric.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 53 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Here are a few examples of the information in the matrix:
The variance of the height variable is 8.74. Thus the standard deviation is√
8.74 = 2.956.
The variance of the left foot measurement is 1.908 (in the 3rd diagonal
element). Thus the standard deviation for this variable is
√
1.908 = 1.381.
The covariance between height and left arm is 3.022, found in the 1st row,
2nd column and also in the 2nd row, 1st column.
The covariance between left foot and left arm is 1.234, found in the 3rd row,
2nd column and also in the 2nd row, 3rd column.
All covariance values are positive so all pairwise associations are positive.
But, the magnitudes do not tell us about the strength of the associations. To
assess the strength of an association, we use correlation values.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 54 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Correlation
This suggests an alternative measure of association. The population
correlation is defined to be equal to the population covariance divided by the
product of the population standard deviations:
ρjk =
σjk
σj σk
.
The population correlation may be estimated by substituting into the formula
the sample covariances and standard deviations:
rjk =
sij
si sj
=
n
i=1 Xij Xik − (
n
i=1 Xij )(
n
i=1 Xik )/n2
(
n
i=1 X2
ij − (
n
i=1 Xij /n)2)(
n
i=1 X2
ik − (
n
i=1 Xik /n)2)
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 55 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Correlation
It is very important to note that the population as well as the sample
correlation must lie between -1 and 1.
−1 ≤ ρjk ≤ 1
−1 ≤ rjk ≤ 1
Therefore:
ρjk = 0 indicates, as you might expect, that the two variables are uncorrelated .
ρjk close to +1 will indicate a strong positive dependence
ρjk close to -1 indicates a strong negative dependence.
Sample correlation coefficients also have similar interpretation.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 56 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Correlation matrix
For a collection of p variables, the correlation matrix is a p × p matrix that
displays the correlations between pairs of variables.
For instance, the value in the j-th row and k-th column gives the correlation
between variables xj and xk .
The correlation matrix is symmetric so that the value in the k-th row and
j-th column is also the correlation between variables xj and xk .
The diagonal elements of the correlation matrix are all identically equal to 1.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 57 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Correlation matrix
The sample correlation matrix is denoted as R.
R =





1 r12 . . . r1p
r21 1 . . . r2p
...
...
...
...
rp1 rp2 . . . 1





Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 58 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
The following covariance matrix show the pairwise covariances for the height,
left forearm, left foot and head circumference measurements of n = 30
female college students.
Here are two examples of calculating a correlation coefficient:
The correlation between height and left forearm is 3.022√
8.74
√
2.402
= 0.66.
The correlation between head circumference and left foot is
0.118√
3.434
√
1.908
= 0.046.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 59 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
The complete sample correlation matrix for this example is the following:
Overall, we see moderately strong linear associations among the variables
height, left arm and left foot and quite weak (almost 0) associations between
head circumference and the other three variables.
In practice, use scatter plots of the variables to fully understand the
associations between variables. It is not a good idea to rely on correlations
without seeing the plots. Correlation values are affected by outliers and
curvilinearity.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 60 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Overall measures of dispersion
Sometimes it is also useful to have an overall measure of dispersion in the
data.
In this measure it would be good to include all of the variables
simultaneously, rather than one at a time.
In the past we looked at the individual variables and their variances to
measure the individual variances.
Here we are going to look at measures of dispersion of all variables together,
particularly we are going to look at such measures that look at total variation.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 61 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Overall measures of dispersion
The variance σ2
j measures the dispersion of an individual variable Xj .
The following two are used to measure the dispersion of all variables
together.
Total Variation
Generalized Variance
To understand total variation we first must find the trace of a square matrix.
A square matrix is a matrix that has an equal number of columns and rows.
Important examples of square matrices include the variance-covariance and
correlation matrices.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 62 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Overall measures of dispersion
The trace of an n × n matrix A is
trace(A) =
n
i=1
aii .
For instance, in a 10 × 10 matrix, the trace is the sum of the diagonal
elements.
The total variation, therefore, of a random vector X is simply the trace of the
population variance-covariance matrix.
trace(Σ) = σ2
1 + σ2
2 + . . . σ2
p.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 63 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Overall measures of dispersion
Thus, the total variation is equal to the sum of the population variances.
The total variation can be estimated by:
trace(S) = s2
1 + s2
2 + · · · + s2
p .
The total variation is of interest for principal components analysis and factor
analysis and we will look at these concepts later in this course.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 64 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Let us use the data from the USDA women’s health survey again to illustrate
this.
We have taken the variances for each of the variables from software output
and have placed them in the table below.
Variable Variance
Calcium 157,829.4
Iron 35.8
Protein 934.9
Vitamin A 2,668,452.4
Vitamin C 5,416.3
Total 2,832,668.8
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 65 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
The total variation for the nutrient intake data is determined by simply
adding up all of the variances for each of the individual variables.
The total variation equals 2,832,668.8. This is a very large number.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 66 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Problem: The problem with total variation is that it does not take into
account correlations among the variables.
These plots show simulated data for pairs of variables with different levels of
correlation. In each case, the variances for both variables are equal to 1, so
that the total variation is 2.
The corresponding variance-covariance matrix is:
Σ =
1 ρjk σj σk
ρjk σj σk 1
where ρjk is the correaltion coefficient between the two variables.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 67 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
When the correlation r = 0, then we see a shotgun-blast pattern of points, widely
dispersed over the entire range of the plot.
Σ =
1 0
0 1
, being the Total Variation trace(Σ) = 2
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 68 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Increasing the correlation to r = 0.7, we see an oval-shaped pattern. Note that
the points are not as widely dispersed.
Σ =
1 0.7
0.7 1
, being the Total Variation trace(Σ) = 2
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 69 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Example
Increasing the correlation to r = 0.9, we see that the points fall along a 45 degree
line, and are even less dispersed.
Σ =
1 0.9
0.9 1
, being the Total Variation trace(Σ) = 2
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 70 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Generalized variance
To take into account the correlations among pairs of variables an alternative
measure of overall variance is suggested.
This measure takes a large value when the various variables show very little
correlation among themselves.
In contrast, this measure takes a small value if the variables show very strong
correlation among themselves, either positive or negative.
This particular measure of dispersion is the generalized variance. In order to
define the generalized variance we first define the determinant of the matrix.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 71 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Generalized variance
We will start simple with a 2 × 2 matrix and then we will move on to more
general definitions for larger matrices.
Let us consider the determinant of a 2 × 2 matrix B as shown below.
Here we can see that it is the product of the two diagonal elements minus the
product of the off-diagonal elements.
|B| =
b11 b12
b21 b22
= b11b22 − b12b21.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 72 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Generalized variance
Here is an example of a simple matrix that has the elements 5, 1, 2 and 4.
You will get the determinant 18.
The product of the diagonal 5 × 4 subtracting the elements of the
off-diagonal 1 × 2 yields an answer of 18:
5 1
2 4
= 5 × 4 − 1 × 2 = 20 − 2 = 18.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 73 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Generalized variance
More generally the determinant of a general p × p matrix B is given by the
expression shown below:
|B| =
p
j=1
(−1)j+1
b1j |B1j |.
The expression involves the sum over all of the first row of B. Note that
these elements are noted by b1j .
These are pre-multiplied by −1 raised to the (j + 1)-th power, so that
basically we are going to have alternating plus and minus signs in our sum.
The matrix B1j is obtained by deleting row i and column j from the matrix B.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 74 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Generalized variance
For a 3 × 3 matrix the determinant is:
|B| =
b11 b12 b13
b21 b22 b23
b31 b32 b33
= b11b22b33 + b12b23b31 + b13b21b32
−b13b22b31 − b32b23b11 − b33b21b12.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 75 / 122
Multivariate data analysis basics Parameters and descriptive statistics
Generalized variance
By definition the generalized variance of a random vector X is equal to |Σ|,
the determinant of the variance-covariance matrix.
The generalized variance can be estimated by calculating |S|, the
determinant of the sample variance-covariance matrix.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 76 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Subsection 3
Matrix manipulations for sample statistics
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 77 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
The Mean and mean corrected data
Suppose we have a n × p data matrix X and a 1 × n vector of ones, 1 . The
mean vector is given by:
¯x =
1
n
1 X,
and the mean corrected data are given by
Xm = X − 1¯x
where Xm gives the matrix containing the mean-corrected data.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 78 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Sum of squares of cross products and covariance
The matrix of sum of squares and cross products (SSCP) SSCPm is given by
SSCPm = XmXm
and the sample variance-covariance matrix is given by
S =
1
n − 1
SSCPm =
1
n − 1
XmXm
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 79 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Standardized data
If we define a p × p diagonal matrix D, which has variances of the variables
in the diagonal, then standardized data are given by
Xs = XmD−1/2
.
The SSCPs of standardized data is given by
SSCPs = XsXs
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 80 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Correlation matrix
The correlation matrix is given by
R =
1
n − 1
SSCPs =
1
n − 1
XsXs
The generalized variance is given by |S|
All these computations can be easily performed in R using the matrix
operations...
The following table presents and hypothetical data set
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 81 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Firm Original Mean-corrected Standardized
data data data
X1 X2 X1 X2 X1 X2
1 13.000 4.000 7.917 3.833 1.619 1.108
2 10.000 6.000 4.917 5.833 1.006 1.686
3 10.000 2.000 4.917 1.833 1.006 0.530
4 8.000 -2.000 2.917 -2.167 0.597 -0.626
5 7.000 4.000 1.917 3.833 0.392 1.108
6 6.000 -3.000 0.917 -3.167 0.187 -0.915
7 5.000 0.000 -0.083 -0.167 -0.017 -0.048
8 4.000 2.000 -1.083 1.833 -0.222 0.530
9 2.000 -1.000 -3.083 -1.167 -0.631 -0.337
10 0.000 -5.000 -5.083 -5.167 -1.040 -1.493
11 -1.000 -1.000 -6.083 -1.167 -1.244 -0.337
12 -3.000 -4.000 -8.083 -4.167 -1.653 -1.204
Mean 5.083 0.167 0.000 0.000 0.000 0.000
SS 262.917 131.667 11.000 11.000
Var 23.902 11.970 23.902 11.970 1.000 1.000
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 82 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Example
The SSCP matrix of this data set is:
SSCP =
262.917 136.375
136.375 131.667
The S matrix is:
S =
1
n − 1
SSCP =
23.902 12.398
12.398 11.970
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 83 / 122
Multivariate data analysis basics Matrix manipulations for sample statistics
Example
The correlation matrix R of these data is:
R =
s2
1
s1s1
s12
s1s2
s21
s2s1
s2
2
s2s2
=
1 0.733
0.733 1
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 84 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Subsection 4
A review of matrix definitions and operations
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 85 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Just a review!
The following material reviews basic matrix definitions and operations.
You should have previous knowledge on matrix algebra.
This review does not relieve you from responsibility of being sufficiently
proficient in these matters!
Therefore be cautious and if you do not feel comfortable with these issues
please get back again to these matters!
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 86 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Transpose of a matrix
The transpose of a matrix A is a matrix in which the rows of the transpose
are the columns of A (so the columns of the transpose are the rows of A).
The transpose of A is denoted as A .
Example: for the 4 × 2 matrix,
A =




3 6
2 5
3 7
1 8




the transpose is the 2 × 4 matrix
A =
3 2 3 1
6 5 7 8
Notice that the columns of A are the rows of A
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 87 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Symmetric matrices
A square matrix A is symmetric if aij = aji , for all i and j.
That is throughout the matrix, the element in the i-th row and j-th column
equals the element in the j-th row and i-th column.
Said another way, a matrix is symmetric when, A = A.
Important examples of symmetric matrices in multivariate statistics include
the variance-covariance matrix and the correlation matrix.
These were defined when we considered descriptive statistics.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 88 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Symmetric matrices
Example: the following is an example of a symmetric matrix.
Notice that the first row and the first column are identical, the second row
and the second column are the same, and that the third row is the same as
the third column. 

4 2 1
2 3 0
1 0 5


Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 89 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Adding two matrices
Two matrices may be added if and only if they have the same dimensions
(same number of rows and also same number of columns as each other. To
add two matrices, add corresponding elements (in terms of location).
Example:
2 4
6 8
+
1 3
5 7
=
2 + 1 4 + 3
6 + 5 8 + 7
=
3 7
11 15
Note: one matrix is subtracted from another in the same way that matrices
are added. To do the subtraction, subtract each element of the first matrix
from the corresponding (in location) element of the first matrix
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 90 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Multiplying a matrix by a scalar
The word scalar is a synonym for a numerical constant.
In matrix terms, a scalar is a matrix with one row and one column.
To multiply a matrix by a scalar, multiply each element in the matrix by the
scalar.
Example: in this example, we multiply a matrix by the value 3.
3 ×
2 4
6 8
=
3 × 2 3 × 4
3 × 6 3 × 8
=
6 12
18 24
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 91 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Multiplication of matrices
To be able to perform the matrix multiplication C = A × B, the number of
columns in the matrix A must equal the number of rows in the matrix B.
The element in the i-th row and j-th column of the answer matrix C, is the
cross product sum of the i-th row of the matrix A and the j-th column of the
matrix B.
This is done of all combinations of rows of A and columns of B.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 92 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Multiplication of matrices
Example: we’ll carry out the multiplication A×B =


1 2
2 3
4 3

×
4 2
1 3
.
The answer is
C =


1 × 4 + 2 × 1 1 × 2 + 2 × 3
2 × 4 + 3 × 1 2 × 2 + 3 × 3
4 × 4 + 3 × 1 4 × 2 + 3 × 3

 =


6 8
11 13
19 17

 .
Notice, for instance, that the element in the 1st row and 1st column of the
answer C is the cross product sum of the 1st row of A and the 1st row of B.
As another example, the 2nd row, 1st column element of C is the cross
product sum of the 2nd row of A and the 1st column of B.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 93 / 122
Multivariate data analysis basics A review of matrix definitions and operations
The Identity Matrix
An identity matrix is a square matrix that has the value one in each main
diagonal position (from upper left to bottom right) and has the value 0 in all
other locations.
As an example of an identity matrix, the 3 × 3 identity matrix is
I =


1 0 0
0 1 0
0 0 1

 .
I is called the identity matrix because multiplication of any square matrix A
by the identity matrix yields the original matrix A as the answer. That is
AI = IA = A.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 94 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Matrix inverse
An inverse (in the traditional sense) can be found only for square matrices.
The inverse of a square matrix A is the matrix A−1
such that
A−1
A = AA−1
= I.
The calculation of an inverse for large matrices is a laborious process that
we’ll leave to the computer.
For 2 × 2 matrices, however, the formula is relatively simple.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 95 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Matrix inverse
For
A =
a11 a12
a21 a22
,
the inverse is
A−1
=
1
a11a22 − a12a21
a22 −a12
−a21 a11
.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 96 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Matrix inverse
Example: we’ll determine the inverse of
A =
10 6
8 5
.
The inverse is
A−1
=
1
10 × 5 − 6 × 8
5 −6
−8 10
=
1
2
5 −6
−8 10
=
2.5 −3
−4 5
.
You might want to check that A−1
A = I. (It does!)
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 97 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition
Let A be a p × p square matrix and I be the p × p identity matrix.
The scalars λ1, λ2, ..., λp satisfying the characteristic equation
|A − λI| = 0
are called the eigenvalues (or characteristic roots) of matrix A.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 98 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition
On the left-hand side, we have the matrix A minus λ times the Identity
matrix.
When we calculate the determinant of the resulting matrix, we end up with a
polynomial of order p.
Setting this polynomial equal to zero, and solving for λ we obtain the desired
eigenvalues.
In general, we will have p solutions and so there are p eigenvalues, not
necessarily all unique.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 99 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example
Let
A =
1 0
1 3
.
Then
|A − λI| =
1 0
1 3
− λ
1 0
0 1
=
1 − λ 0
1 3 − λ
= (1 − λ)(3 − λ)
implies that there are two roots, λ1 = 1 and λ2 = 3. The eigenvalues of A
are 3 and 1 (the eigenvalues are in generally listed in descending order).
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 100 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition
The corresponding eigenvectors e1, e2, ..., ep are obtained by solving the
expression below:
(A − λj I)ej = 0
Here, we have the difference between the matrix A minus the j-th eigenvalue
times the Identity matrix, this quantity is then multiplied by the j-th
eigenvector and set it all equal to zero.
This will obtain the eigenvector ej associated with eigenvalue λj .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 101 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition
Note: This does not have a generally have a unique solution.
So, to obtain a unique solution we will often require that ej transposed ej is
equal to 1.
Or, if you like, the sum of the square elements of ej is equal to 1, which
means the length (norm) of the eigenvector is the unity.
ej ej = 1
Also note that eigenvectors corresponding to different eigenvalues are
orthogonal. In situations, where two (or more) eigenvalues are equal,
corresponding eigenvectors will still be orthogonal.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 102 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example (continued)
The eigenvectors of matrix A associated to those eigenvalues (1 and 3) can
be determined by solving the following equations:
Ax = λ1x
1 0
1 3
x1
x2
= 1 ×
x1
x2
x1
x1 + 3x2
=
x1
x2
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 103 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example (continued)
There are many solutions! Therefore, setting x2 = 1 (arbitrarily) gives
x1 = −2, and hence
−2
1
is an eigenvector corresponding to the eigenvalue 1.
After transforming for unit norm:
e = (−2/
√
5 1/
√
5).
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 104 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example (continued)
Ax = λ2x
1 0
1 3
x1
x2
= 3 ×
x1
x2
x1
x1 + 3x2
=
3x1
3x2
Again, setting x2 = 1 and x1 = 0
0
1
is an eigenvector corresponding to the eigenvalue 3.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 105 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition
Any symmetric square matrix can be reconstructed from its eigenvalues and
eigenvectors.
Let A be a p × p symmetric matrix. Then A can be expressed in terms of its
p eigenvalue-eigenvector pairs (λi , ei ) as
A =
p
j=1
λj ej ej
Or, if you dispose the eigenvalues in a diagonal matrix Λ and the eigenvectors
in a matrix E, then
A = EΛE .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 106 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example
For example, let
A =
2.2 0.4
0.4 2.8
Then
|A − λI| = λ2
− 5λ + 6.16 − 0.16 = (λ − 3)(λ − 2)
so A has eigenvalues λ1 = 3 and λ2 = 2.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 107 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example
The corresponding eigenvectors are e1 = ( 1/
√
5 2/
√
5 ) and
e2 = ( −2/
√
5 1/
√
5 ), respectively.
Therefore
A =
2.2 0.4
0.4 2.8
= 3
1/
√
5
2/
√
5
1/
√
5 2/
√
5 +
+2
−2/
√
5
1/
√
5
−2/
√
5 1/
√
5
=
0.6 1.2
1.2 2.4
+
1.6 −0.8
−0.8 0.4
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 108 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Example
Or,
A =
2.2 0.4
0.4 2.8
=
=
1/
√
5 −2/
√
5
2/
√
5 1/
√
5
3 0
0 2
1/
√
5 2/
√
5
−2/
√
5 1/
√
5
=
3/
√
5 −4/
√
5
6/
√
5 2/
√
5
1/
√
5 −2/
√
5
2/
√
5 1/
√
5
=
3/5 + 8/5 6/5 − 4/5
6/5 − 4/5 12/5 + 2/5/
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 109 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition
Usually the matrix A is taken to be either the variance-covariance matrix Σ,
or the correlation matrix, or their estimates S and R, respectively.
Eigenvalues and eigenvectors are used for:
Computing prediction and confidence ellipses
Principal Components Analysis (later in the course)
Factor Analysis (also later in this course)
Discriminant analysis (also later in this course)
Multidimensional scaling (also later in this course)
...
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 110 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Consider the correlation matrix R as shown below:
R =
1 ρ
ρ 1
.
Then, using the definition of the eigenvalues, we must calculate the
determinant of R − λ times the Identity matrix.
|R − λI| =
1 ρ
ρ 1
− λ
1 0
0 1
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 111 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Carrying out the math we end up with the matrix with 1 − λ on the diagonal
and ρ on the off-diagonal. Then calculating this determinant we obtain 1 − λ
squared minus ρ2
:
1 − λ ρ
ρ 1 − λ
= (1 − λ)2
− ρ2
= λ2
− 2λ + 1 − ρ2
.
Setting this expression equal to zero we end up with the following...
λ2
− 2λ + 1 − ρ2
= 0.
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 112 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
To solve for λ we use the general result that any solution to the second order
polynomial below:
ay2
+ by + c = 0
is given by the following expression:
y =
−b ±
√
b2 − 4ac
2a
.
Here, a = 1, b = −2 (the term that precedes λ and c is equal to 1 − ρ2
.
Substituting these terms in the equation above, we obtain that λ must be
equal to 1 plus or minus the correlation ρ.
λ =
2 ± 22 − 4(1 − ρ2)
2
= 1 ± 1 − (1 − ρ2) = 1 ± ρ
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 113 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
We will take the following solutions:
λ1 = 1 + ρ
λ2 = 1 − ρ
Next, to obtain the corresponding eigenvectors, we must solve a system of
equations below:
(R − λ1I)e1 = 0
and
(R − λ2I)e2 = 0
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 114 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Or in other words, this is translated for this specific problem in the expression
below:
1 ρ
ρ 1
− λ1
1 0
0 1
e11
e12
=
0
0
This simplifies as follows:
1 − λ1 ρ
ρ 1 − λ1
e11
e12
=
0
0
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 115 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Yielding a system of two equations with two unknowns:
(1 − λ1)e11 + ρe12 = 0
ρe11 + (1 − λ1)e12 = 0
Note that this does not have a unique solution. If (e11, e12) is one solution,
then a second solution can be obtained by multiplying the first solution by
any non-zero constant c, i.e., (ce11, ce12). Therefore, we will require the
additional condition that the sum of the squared values of e11 and e12 are
equal to 1.
e2
11 + e2
12 = 1
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 116 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Consider the first equation:
(1 − λ1)e11 + ρe12 = 0
Solving this equation for e12 and we obtain the following:
e12 = −
(1 − λ1)
ρ
e11
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 117 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Substituting this into e2
11 + e2
12 = 1 we get the following:
e2
11 +
(1 − λ1)2
ρ2
e2
11 = 1
Recall that λ = 1 ± ρ. In either case we end up finding that (1 − λ1)2
= ρ2
,
so that the expression above simplifies to:
2e2
11 = 1
Or, in other words:
e11 =
1
√
2
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 118 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Using the expression for e12 which we obtained above,
e12 = −
1 − λ1
ρ
e11 =
1
√
2
we get
e11 =
1
√
2
and e12 =
1
√
2
for λ1 = 1 + ρ
Using the same reasoning we get
e21 =
1
√
2
and e22 = −
1
√
2
for λ2 = 1 − ρ
Therefore, the two eigenvectors are given by the two vectors as shown below:
1√
2
1√
2
for λ = 1 + ρ and
1√
2
− 1√
2
for λ = 1 − ρ
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 119 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Some properties of the eigenvalues of the variance-covariance matrix are to
be considered at this point. Suppose that λ1, λ2, ..., λp are the eigenvalues of
the variance-covariance matrix Σ.
By definition, the total variation is given by the sum of the variances. It turns
out that this is also equal to the sum of the eigenvalues of the
variance-covariance matrix. Thus, the total variation is:
p
j=1
s2
j = s2
1 + s2
2 + · · · + s2
p = λ1 + λ2 + · · · + λp =
p
j=1
λj
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 120 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Eigenvalues, eigenvectors and spectral decomposition: a
special case
Some properties of the eigenvalues of the variance-covariance matrix are to
be considered at this point. Suppose that λ1, λ2, ..., λp are the eigenvalues of
the variance-covariance matrix Σ.
The generalized variance is equal to the product of the eigenvalues:
|Σ| =
p
j=1
λj = λ1 × λ2 × · · · × λp
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 121 / 122
Multivariate data analysis basics A review of matrix definitions and operations
Quadratic forms
A quadratic form Q(x) in the p variables x1, x2, ...xp is Q(x) = x Ax, where
x = [x1, x2, ..., xp] and A is a p × p symmetric matrix.
Note that a quadratic form can be written as Q(x) =
p
i=1
p
j=1
aij xi xj . For
example,
Q(x) = x1 x2
1 1
1 1
x1
x2
= x2
1 + x1x2 + x2
2 .
Q(x) = x1 x2 x3


1 3 0
3 −1 −2
0 −2 2




x1
x2
x3


= x2
1 + 6x1x2 − x2
2 − 4x2x3 + 2x2
3 .
Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 122 / 122

More Related Content

What's hot (20)

Seminar on Robust Regression Methods
Seminar on Robust Regression MethodsSeminar on Robust Regression Methods
Seminar on Robust Regression Methods
 
In Anova
In  AnovaIn  Anova
In Anova
 
Inferential statistics
Inferential statisticsInferential statistics
Inferential statistics
 
Frequency Distributions
Frequency DistributionsFrequency Distributions
Frequency Distributions
 
Student T - test
Student T -  testStudent T -  test
Student T - test
 
T12 non-parametric tests
T12 non-parametric testsT12 non-parametric tests
T12 non-parametric tests
 
t test
t testt test
t test
 
Posthoc
PosthocPosthoc
Posthoc
 
Friedman test Stat
Friedman test Stat Friedman test Stat
Friedman test Stat
 
Confidence Intervals
Confidence IntervalsConfidence Intervals
Confidence Intervals
 
Measures of dispersion
Measures  of  dispersionMeasures  of  dispersion
Measures of dispersion
 
Diagnostic in poisson regression models
Diagnostic in poisson regression modelsDiagnostic in poisson regression models
Diagnostic in poisson regression models
 
Non Parametric Tests
Non Parametric TestsNon Parametric Tests
Non Parametric Tests
 
The kolmogorov smirnov test
The kolmogorov smirnov testThe kolmogorov smirnov test
The kolmogorov smirnov test
 
Hypothesis testing chi square goodness of fit test
Hypothesis testing chi square goodness of fit testHypothesis testing chi square goodness of fit test
Hypothesis testing chi square goodness of fit test
 
Analysis of variance (ANOVA)
Analysis of variance (ANOVA)Analysis of variance (ANOVA)
Analysis of variance (ANOVA)
 
Anova and T-Test
Anova and T-TestAnova and T-Test
Anova and T-Test
 
Survival analysis
Survival  analysisSurvival  analysis
Survival analysis
 
Experimental Studies
Experimental StudiesExperimental Studies
Experimental Studies
 
Stat topics
Stat topicsStat topics
Stat topics
 

Viewers also liked

แบบเสนอโครงร่างโครงงานคอมพิวเตอร์
แบบเสนอโครงร่างโครงงานคอมพิวเตอร์แบบเสนอโครงร่างโครงงานคอมพิวเตอร์
แบบเสนอโครงร่างโครงงานคอมพิวเตอร์ice mewika
 
Hub and Spoke Presentation
Hub and Spoke Presentation Hub and Spoke Presentation
Hub and Spoke Presentation aliciacorbin
 
Aims Tutorials Pvt. Ltd.
Aims Tutorials Pvt. Ltd.Aims Tutorials Pvt. Ltd.
Aims Tutorials Pvt. Ltd.Aims Tutorials
 
Running open source PHP applications on you IBM i
Running open source PHP applications on you IBM iRunning open source PHP applications on you IBM i
Running open source PHP applications on you IBM iProximity Group
 
PHP Installed on IBM i - the Nickel Tour
PHP Installed on IBM i - the Nickel TourPHP Installed on IBM i - the Nickel Tour
PHP Installed on IBM i - the Nickel TourRod Flohr
 
Ethics In Education Shankara Narayanan
Ethics In Education Shankara NarayananEthics In Education Shankara Narayanan
Ethics In Education Shankara NarayananShankara Narayanan
 
Reaction Injection Molding Tooling
Reaction Injection Molding ToolingReaction Injection Molding Tooling
Reaction Injection Molding ToolingAndrew Hollcraft
 
Was_there_room_for_monetary_discretion_under_the_gold_standard
Was_there_room_for_monetary_discretion_under_the_gold_standardWas_there_room_for_monetary_discretion_under_the_gold_standard
Was_there_room_for_monetary_discretion_under_the_gold_standardJohn Phelan
 

Viewers also liked (15)

Турбюро «Компас»
Турбюро «Компас»Турбюро «Компас»
Турбюро «Компас»
 
แบบเสนอโครงร่างโครงงานคอมพิวเตอร์
แบบเสนอโครงร่างโครงงานคอมพิวเตอร์แบบเสนอโครงร่างโครงงานคอมพิวเตอร์
แบบเสนอโครงร่างโครงงานคอมพิวเตอร์
 
2015 09 23_prezentatsia
2015 09 23_prezentatsia2015 09 23_prezentatsia
2015 09 23_prezentatsia
 
Hub and Spoke Presentation
Hub and Spoke Presentation Hub and Spoke Presentation
Hub and Spoke Presentation
 
Keeping up with PHP
Keeping up with PHPKeeping up with PHP
Keeping up with PHP
 
Lean Six SIgma Green Belt
Lean Six SIgma Green BeltLean Six SIgma Green Belt
Lean Six SIgma Green Belt
 
The new features of PHP 7
The new features of PHP 7The new features of PHP 7
The new features of PHP 7
 
Aims Tutorials Pvt. Ltd.
Aims Tutorials Pvt. Ltd.Aims Tutorials Pvt. Ltd.
Aims Tutorials Pvt. Ltd.
 
Law Diplom
Law DiplomLaw Diplom
Law Diplom
 
Running open source PHP applications on you IBM i
Running open source PHP applications on you IBM iRunning open source PHP applications on you IBM i
Running open source PHP applications on you IBM i
 
Getting started with PHP on IBM i
Getting started with PHP on IBM iGetting started with PHP on IBM i
Getting started with PHP on IBM i
 
PHP Installed on IBM i - the Nickel Tour
PHP Installed on IBM i - the Nickel TourPHP Installed on IBM i - the Nickel Tour
PHP Installed on IBM i - the Nickel Tour
 
Ethics In Education Shankara Narayanan
Ethics In Education Shankara NarayananEthics In Education Shankara Narayanan
Ethics In Education Shankara Narayanan
 
Reaction Injection Molding Tooling
Reaction Injection Molding ToolingReaction Injection Molding Tooling
Reaction Injection Molding Tooling
 
Was_there_room_for_monetary_discretion_under_the_gold_standard
Was_there_room_for_monetary_discretion_under_the_gold_standardWas_there_room_for_monetary_discretion_under_the_gold_standard
Was_there_room_for_monetary_discretion_under_the_gold_standard
 

Similar to Lecture 1

SPSS statistics - get help using SPSS
SPSS statistics - get help using SPSSSPSS statistics - get help using SPSS
SPSS statistics - get help using SPSScsula its training
 
Level of Measurement, Frequency Distribution,Stem & Leaf
Level of Measurement, Frequency Distribution,Stem & Leaf   Level of Measurement, Frequency Distribution,Stem & Leaf
Level of Measurement, Frequency Distribution,Stem & Leaf Qasim Raza
 
MELJUN CORTES research lectures_evaluating_data_statistical_treatment
MELJUN CORTES research lectures_evaluating_data_statistical_treatmentMELJUN CORTES research lectures_evaluating_data_statistical_treatment
MELJUN CORTES research lectures_evaluating_data_statistical_treatmentMELJUN CORTES
 
IoT Lecture MaterialLecture Material.pdf
IoT Lecture MaterialLecture Material.pdfIoT Lecture MaterialLecture Material.pdf
IoT Lecture MaterialLecture Material.pdfsaimawarsi
 
Statistics and types of statistics .docx
Statistics and types of statistics .docxStatistics and types of statistics .docx
Statistics and types of statistics .docxHwre Idrees
 
STATISTICAL PROCEDURES (Discriptive Statistics).pptx
STATISTICAL PROCEDURES (Discriptive Statistics).pptxSTATISTICAL PROCEDURES (Discriptive Statistics).pptx
STATISTICAL PROCEDURES (Discriptive Statistics).pptxMuhammadNafees42
 
applied multivariate statistical techniques in agriculture and plant science 2
applied multivariate statistical techniques in agriculture and plant science 2applied multivariate statistical techniques in agriculture and plant science 2
applied multivariate statistical techniques in agriculture and plant science 2amir rahmani
 
CABT Math 8 measures of central tendency and dispersion
CABT Math 8   measures of central tendency and dispersionCABT Math 8   measures of central tendency and dispersion
CABT Math 8 measures of central tendency and dispersionGilbert Joseph Abueg
 
Spsshelp 100608163328-phpapp01
Spsshelp 100608163328-phpapp01Spsshelp 100608163328-phpapp01
Spsshelp 100608163328-phpapp01Henock Beyene
 
Introduction to Biostatistics_20_4_17.ppt
Introduction to Biostatistics_20_4_17.pptIntroduction to Biostatistics_20_4_17.ppt
Introduction to Biostatistics_20_4_17.pptnyakundi340
 
Data analysis
Data analysisData analysis
Data analysisneha147
 
Data Analysis and Statistics
Data Analysis and StatisticsData Analysis and Statistics
Data Analysis and StatisticsT.S. Lim
 
KIT-601 Lecture Notes-UNIT-2.pdf
KIT-601 Lecture Notes-UNIT-2.pdfKIT-601 Lecture Notes-UNIT-2.pdf
KIT-601 Lecture Notes-UNIT-2.pdfDr. Radhey Shyam
 
Stastistics in Physical Education - SMK.pptx
Stastistics in Physical Education - SMK.pptxStastistics in Physical Education - SMK.pptx
Stastistics in Physical Education - SMK.pptxshatrunjaykote
 

Similar to Lecture 1 (20)

Data Analysis
Data Analysis Data Analysis
Data Analysis
 
SPSS statistics - get help using SPSS
SPSS statistics - get help using SPSSSPSS statistics - get help using SPSS
SPSS statistics - get help using SPSS
 
Level of Measurement, Frequency Distribution,Stem & Leaf
Level of Measurement, Frequency Distribution,Stem & Leaf   Level of Measurement, Frequency Distribution,Stem & Leaf
Level of Measurement, Frequency Distribution,Stem & Leaf
 
MELJUN CORTES research lectures_evaluating_data_statistical_treatment
MELJUN CORTES research lectures_evaluating_data_statistical_treatmentMELJUN CORTES research lectures_evaluating_data_statistical_treatment
MELJUN CORTES research lectures_evaluating_data_statistical_treatment
 
IoT Lecture MaterialLecture Material.pdf
IoT Lecture MaterialLecture Material.pdfIoT Lecture MaterialLecture Material.pdf
IoT Lecture MaterialLecture Material.pdf
 
Statistics and types of statistics .docx
Statistics and types of statistics .docxStatistics and types of statistics .docx
Statistics and types of statistics .docx
 
STATISTICAL PROCEDURES (Discriptive Statistics).pptx
STATISTICAL PROCEDURES (Discriptive Statistics).pptxSTATISTICAL PROCEDURES (Discriptive Statistics).pptx
STATISTICAL PROCEDURES (Discriptive Statistics).pptx
 
applied multivariate statistical techniques in agriculture and plant science 2
applied multivariate statistical techniques in agriculture and plant science 2applied multivariate statistical techniques in agriculture and plant science 2
applied multivariate statistical techniques in agriculture and plant science 2
 
Presentation of BRM.pptx
Presentation of BRM.pptxPresentation of BRM.pptx
Presentation of BRM.pptx
 
Introduction to spss
Introduction to spssIntroduction to spss
Introduction to spss
 
CABT Math 8 measures of central tendency and dispersion
CABT Math 8   measures of central tendency and dispersionCABT Math 8   measures of central tendency and dispersion
CABT Math 8 measures of central tendency and dispersion
 
Spsshelp 100608163328-phpapp01
Spsshelp 100608163328-phpapp01Spsshelp 100608163328-phpapp01
Spsshelp 100608163328-phpapp01
 
Introduction to Biostatistics_20_4_17.ppt
Introduction to Biostatistics_20_4_17.pptIntroduction to Biostatistics_20_4_17.ppt
Introduction to Biostatistics_20_4_17.ppt
 
Data analysis
Data analysisData analysis
Data analysis
 
Data analysis
Data analysisData analysis
Data analysis
 
Data Analysis and Statistics
Data Analysis and StatisticsData Analysis and Statistics
Data Analysis and Statistics
 
Lessons learnt in statistics essay
Lessons learnt in statistics essayLessons learnt in statistics essay
Lessons learnt in statistics essay
 
KIT-601 Lecture Notes-UNIT-2.pdf
KIT-601 Lecture Notes-UNIT-2.pdfKIT-601 Lecture Notes-UNIT-2.pdf
KIT-601 Lecture Notes-UNIT-2.pdf
 
Stastistics in Physical Education - SMK.pptx
Stastistics in Physical Education - SMK.pptxStastistics in Physical Education - SMK.pptx
Stastistics in Physical Education - SMK.pptx
 
Lesson 6 chapter 4
Lesson 6   chapter 4Lesson 6   chapter 4
Lesson 6 chapter 4
 

Recently uploaded

Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...D. B. S. College Kanpur
 
FREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by naFREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by naJASISJULIANOELYNV
 
Neurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trNeurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trssuser06f238
 
OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024innovationoecd
 
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptxTHE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptxNandakishor Bhaurao Deshmukh
 
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRCall Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRlizamodels9
 
《Queensland毕业文凭-昆士兰大学毕业证成绩单》
《Queensland毕业文凭-昆士兰大学毕业证成绩单》《Queensland毕业文凭-昆士兰大学毕业证成绩单》
《Queensland毕业文凭-昆士兰大学毕业证成绩单》rnrncn29
 
Harmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms PresentationHarmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms Presentationtahreemzahra82
 
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)riyaescorts54
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxBerniceCayabyab1
 
Dubai Calls Girl Lisa O525547819 Lexi Call Girls In Dubai
Dubai Calls Girl Lisa O525547819 Lexi Call Girls In DubaiDubai Calls Girl Lisa O525547819 Lexi Call Girls In Dubai
Dubai Calls Girl Lisa O525547819 Lexi Call Girls In Dubaikojalkojal131
 
User Guide: Orion™ Weather Station (Columbia Weather Systems)
User Guide: Orion™ Weather Station (Columbia Weather Systems)User Guide: Orion™ Weather Station (Columbia Weather Systems)
User Guide: Orion™ Weather Station (Columbia Weather Systems)Columbia Weather Systems
 
Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)Tamer Koksalan, PhD
 
Behavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdfBehavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdfSELF-EXPLANATORY
 
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPirithiRaju
 
User Guide: Magellan MX™ Weather Station
User Guide: Magellan MX™ Weather StationUser Guide: Magellan MX™ Weather Station
User Guide: Magellan MX™ Weather StationColumbia Weather Systems
 
Base editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editingBase editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editingNetHelix
 
Pests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdfPests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdfPirithiRaju
 
Citronella presentation SlideShare mani upadhyay
Citronella presentation SlideShare mani upadhyayCitronella presentation SlideShare mani upadhyay
Citronella presentation SlideShare mani upadhyayupadhyaymani499
 

Recently uploaded (20)

Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
 
FREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by naFREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by na
 
Neurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 trNeurodevelopmental disorders according to the dsm 5 tr
Neurodevelopmental disorders according to the dsm 5 tr
 
OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024
 
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptxTHE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
 
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRCall Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
 
《Queensland毕业文凭-昆士兰大学毕业证成绩单》
《Queensland毕业文凭-昆士兰大学毕业证成绩单》《Queensland毕业文凭-昆士兰大学毕业证成绩单》
《Queensland毕业文凭-昆士兰大学毕业证成绩单》
 
Harmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms PresentationHarmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms Presentation
 
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
 
Dubai Calls Girl Lisa O525547819 Lexi Call Girls In Dubai
Dubai Calls Girl Lisa O525547819 Lexi Call Girls In DubaiDubai Calls Girl Lisa O525547819 Lexi Call Girls In Dubai
Dubai Calls Girl Lisa O525547819 Lexi Call Girls In Dubai
 
User Guide: Orion™ Weather Station (Columbia Weather Systems)
User Guide: Orion™ Weather Station (Columbia Weather Systems)User Guide: Orion™ Weather Station (Columbia Weather Systems)
User Guide: Orion™ Weather Station (Columbia Weather Systems)
 
Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)Carbon Dioxide Capture and Storage (CSS)
Carbon Dioxide Capture and Storage (CSS)
 
Behavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdfBehavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdf
 
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
 
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
 
User Guide: Magellan MX™ Weather Station
User Guide: Magellan MX™ Weather StationUser Guide: Magellan MX™ Weather Station
User Guide: Magellan MX™ Weather Station
 
Base editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editingBase editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editing
 
Pests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdfPests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdf
 
Citronella presentation SlideShare mani upadhyay
Citronella presentation SlideShare mani upadhyayCitronella presentation SlideShare mani upadhyay
Citronella presentation SlideShare mani upadhyay
 

Lecture 1

  • 1. Some important information on the course Lecturer: Jorge M. Mendes (jmm@novaims.unl.pt) Lectures timetable: Tuesdays and Thursdays, 3:30pm to 4:15pm Assessment rules: check syllabus Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 1 / 122
  • 2. Some important information on the course Textbooks Everitt, B. and Hothorn, T. (2011). An Introduction to Applied Multivariate Analysis with R, Springer. Johnson, R.A and Wichern (2007). D. W., Applied Multivariate Statistical Analysis, 6th edition, Pearson Prentice Hall. online tools and references: https://www.datacamp.com online tools and references: http://www.statmethods.net And... Reis, E. (1997), Estat´ıstica Multivariada Aplicada, Edi¸c˜oes S´ılabo Sharma, S. (1996). Applied Multivariate Techniques, John Wiley & Sons Timm, N. H. (2002). Applied Multivariate Analysis, Springer Other things: lecture slides, data files, etc. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 2 / 122
  • 3. Some important information on the course Course contents Data analysis basics: organisation of multivariate data; parameters and descriptive statistics; matrix manipulations for sample statistics; review of matrix definitions and operations R tutorial; fundamentals on data manipulation on R Graphical display of multivariate data; Linear combinations of random variables The multivariate normal distribution Principal components analysis Factor analysis Discriminant analysis Cluster analysis Canonical correlation analysis Multidimensional scaling ... eventually something else! Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 3 / 122
  • 4. Lecture 1 Multivariate Data Analysis Basics Jorge M. Mendes Nova Information Management School Universidade Nova de Lisboa 2015 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 4 / 122
  • 5. First things first This first lesson will introduce the basic tools to deal with multivariate data and multivariate techniques. Topics covered include multivariate sample statistics and spectral decomposition of a matrix. After successfully completing this lesson, you should be able to: Understand the notation used in multivariate analysis; Understand the organization of multivariate data using vector and matrix notation; Perform basic operations on matrices and vectors; Interpret measures of central tendency, dispersion, and association; Calculate sample means, variances, covariances, and correlations using a hand calculator; Compute the spectral decomposition of a matrix (eigenvalues, eigenvectors, etc.) Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 5 / 122
  • 6. Multivariate data analysis basics Organisation of multivariate data Subsection 1 Organisation of multivariate data Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 6 / 122
  • 7. Multivariate data analysis basics Organisation of multivariate data Introduction Univariate statistics is concerned with random scalar variable Y . In multivariate analysis, we are concerned with the joint analysis of multiple dependent variables. These variables can be represented using matrices and vectors. This provides simplification of notation and a format for expressing important formulas. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 7 / 122
  • 8. Multivariate data analysis basics Organisation of multivariate data Example Suppose that we measure the variables x1 = height (cm), x2 = left forearm length (cm) and x3 = left foot length for participants in a study of the physical characteristics of adult humans. These three variables can be represented in the following column vector: x =   x1 x2 x3   Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 8 / 122
  • 9. Multivariate data analysis basics Organisation of multivariate data Example The observed data for a specific individual, say the i-th individual, might also be represented in an analogous vector. Suppose that the i-th person in the sample has height = 175 cm, forearm length = 25.5 cm and foot length = 27 cm. In vector notation these observed data could be written as: xi =   xi1 xi2 xi3   =   175 25.5 27.0   Notice the use and placement of the subscript i to represent the i-th individual. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 9 / 122
  • 10. Multivariate data analysis basics Organisation of multivariate data Definitions of matrix and vector A matrix is two-dimensional array of numbers of formulas. A vector is a matrix with either only one column or only one row. A column vector has only one column. A row vector has only one row. The dimension of a matrix is expressed as number of rows × number of columns. For instance, a matrix with 10 rows and 3 columns is said to be a 10 × 3 matrix. The vectors written in previous Example are 3 × 1 matrices. A square matrix is one for which the numbers of rows and columns are the same. For instance, a 4 × 4 matrix is a square matrix. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 10 / 122
  • 11. Multivariate data analysis basics Organisation of multivariate data The data matrix in multivariate problems Usually the observed data are represented by a matrix in which the rows are observations and the columns are variables. This is exactly the way the data are normally prepared for statistical software. The usual notation is n: the number of observed units (people, animals, companies, etc.) p: number of variables measured on each unit. Thus the data matrix will be an n × p matrix. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 11 / 122
  • 12. Multivariate data analysis basics Organisation of multivariate data Example Suppose that we have scores for n = 6 college students who have taken the verbal and the science subtests of a College Qualification test (CQT). We have p = 2 variables: (1) the verbal score and (2) the science score for each student. The data matrix is the following 6 × 2 matrix: X =         41 26 39 26 53 21 67 33 61 27 67 29         Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 12 / 122
  • 13. Multivariate data analysis basics Organisation of multivariate data Example In the matrix just given, the first column gives the data for x1 = verbal score whereas the second column gives data for x2 = science score. Each row gives data for a student in the sample. To repeat - the rows are observations, the columns are variables. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 13 / 122
  • 14. Multivariate data analysis basics Organisation of multivariate data Notation notes Note that we have used a small x to denote the vector of variables in first example and a large X to represent the data matrix in second example. It should also be noted that, in matrix terms, the i-th row in the data matrix X is the transpose of the data vector xi = xi1 xi2 , as we defined data vectors in first Example. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 14 / 122
  • 15. Multivariate data analysis basics Parameters and descriptive statistics Subsection 2 Parameters and descriptive statistics Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 15 / 122
  • 16. Multivariate data analysis basics Parameters and descriptive statistics Introduction Three aspects of the data are of importance, the first two of which you should already be familiar with from univariate statistics. These are: Central Tendency. What is a typical value for each variable? Dispersion. How far apart are the individual observations from a central value for a given variable? Association. This might (or might not!) be a new measure for you. When more than one variable are studied together how does each variable relate to the remaining variables? How are the variables simultaneously related to one another? Are they positively or negatively related? Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 16 / 122
  • 17. Multivariate data analysis basics Parameters and descriptive statistics Population parameters and sample statistics Statistics, as a subject matter, is the science and art of using sample information to make generalizations about populations. A population is the collection of all people, plants, animals, or objects of interest about which we wish to make statistical inferences (generalizations). A population parameter is a numerical characteristic of a population. In nearly all statistical problems we do not know the value of a parameter because we do not measure the entire population. We use sample data to make an inference about the value of a parameter. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 17 / 122
  • 18. Multivariate data analysis basics Parameters and descriptive statistics Population parameters and sample statistics A sample is the subset of the population that we actually measure or observe. A sample statistic is a numerical characteristic of a sample. A sample statistic estimates the unknown value of a population parameter. Information collected from sample statistic is sometimes referred to as descriptive statistic. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 18 / 122
  • 19. Multivariate data analysis basics Parameters and descriptive statistics Notation Here are the notation that will be used: Xij : Observation for variable j in subject i. p: Number of variables n: Number of subjects Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 19 / 122
  • 20. Multivariate data analysis basics Parameters and descriptive statistics Example: USDA Women’s Health Survey Let us take a look at an example. In 1985, the USDA commissioned a study of women’s nutrition. Nutrient intake was measured for a random sample of 737 women aged 25-50 years. The following variables were measured: Calcium (mg) Iron (mg) Protein (g) Vitamin A (µg) Vitamin C (mg) Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 20 / 122
  • 21. Multivariate data analysis basics Parameters and descriptive statistics Example: USDA Women’s Health Survey As a notation example let’s use the women’s nutrition data. p = 5 nutritional outcomes n = 737 subjects Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 21 / 122
  • 22. Multivariate data analysis basics Parameters and descriptive statistics Example: USDA Women’s Health Survey In multivariate statistics we will always be working with vectors of observations. So in this case we are going to arrange the data for the p variables on each subject into a vector. In the expression below, Xi is the vector of observations for the i-th subject, i = 1, ..., n(= 737). Therefore, the data for the j-th variable will be located in the j-th element of this subject’s vector, j = 1, ..., p(= 5). Xi =      Xi1 Xi2 ... Xip      Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 22 / 122
  • 23. Multivariate data analysis basics Parameters and descriptive statistics The mean Throughout this course, we’ll use the ordinary notations for the mean of a variable. That is, the symbol µ is used to represent a (theoretical) population mean and the symbol ¯x is used to represent a sample mean computed from observed data. In the multivariate setting, we add subscripts to these symbols to indicate the specific variable for which the mean is being given. For instance, µ1 represents the population mean for variable x1 and ¯x1 denotes a sample mean based on observed data for variable x1. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 23 / 122
  • 24. Multivariate data analysis basics Parameters and descriptive statistics The mean The population mean is the measure of central tendency for the population. Here, the population mean for variable j is µj = E(Xij ) The notation E stands for statistical expectation; here E(Xij ) is the mean of Xij over all members of the population, or equivalently, over all random draws from a stochastic model. For example, µj = E(Xij ) may be the mean of a normal variable. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 24 / 122
  • 25. Multivariate data analysis basics Parameters and descriptive statistics The mean The population mean µj for variable j can be estimated by the sample mean ¯xj = 1 n n i=1 Xij Note: the sample mean ¯xj , because it is a function of our random data is also going to have a mean itself. In fact, the population mean of the sample mean is equal to population mean µj ; i.e., E(¯xj ) = µj Therefore, the ¯xj is unbiased for µj . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 25 / 122
  • 26. Multivariate data analysis basics Parameters and descriptive statistics The mean Another way of saying this is that the mean of the ¯xj ’s over all possible samples of size n is equal to µj . Recall that the population mean vector is µ which is a collection of the means for each of the population means for each of the different variables. µ =      µ1 µ2 ... µp      Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 26 / 122
  • 27. Multivariate data analysis basics Parameters and descriptive statistics The mean We can estimate this population mean vector, µ, by ¯x. This is obtained by collecting the sample means from each of the variables in a single vector. This is shown below. ¯x =      ¯x1 ¯x2 ... ¯xp      =      1 n n i=1 Xi1 1 n n i=1 Xi2 ... 1 n n i=1 Xip      = 1 n n i=1 Xi Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 27 / 122
  • 28. Multivariate data analysis basics Parameters and descriptive statistics The mean Just as the sample means, ¯xj , for the individual variables are unbiased for their respective population means, note that the sample mean vector is unbiased for the population mean vector. E(¯x) = E      ¯x1 ¯x2 ... ¯xp      =      E(¯x1) E(¯x2) ... E(¯xp)      =      µ1 µ2 ... µp      = µ Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 28 / 122
  • 29. Multivariate data analysis basics Parameters and descriptive statistics Dispersion: variance and standard deviation A variance measures the degree of spread (dispersion) in a variable’s values. Theoretically, a population variance is the average squared difference between a variable’s values and the mean for that variable. The population variance for variable xj is σ2 j = E(xj − µj )2 Note that the squared residual (xj − µj )2 is a function of the random variable Xij . Therefore, the squared residual itself is random, and has a population mean. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 29 / 122
  • 30. Multivariate data analysis basics Parameters and descriptive statistics Dispersion: variance and standard deviation The population variance is thus the population mean of the squared residual. We see that if the data tend to be far away from the mean, the squared residual will tend to be large, and hence the population variance will also be large. Conversely, if the the data tend to be close to the mean, the squared residual will tend to be small, and hence the population variance will also be small. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 30 / 122
  • 31. Multivariate data analysis basics Parameters and descriptive statistics Dispersion: variance and standard deviation The population variance σ2 j can be estimated by the sample variance s2 j = 1 n − 1 n i=1 (Xij − ¯xj )2 = 1 n − 1 n i=1 X2 ij − n n − 1 ¯x2 j The first expression in this formula is most suitable for interpreting the sample variance. We see that it is a function of the squared residuals; that is, take difference between the individual observations and their sample mean, and then square the result. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 31 / 122
  • 32. Multivariate data analysis basics Parameters and descriptive statistics Dispersion: variance and standard deviation Here, we may observe that if tend to be far away from their sample means, then the squared residuals and hence the sample variance will also tend to be large. If on the other hand, if the observations tend to be close to their respective sample means, then the squared differences between the data and their means will be small, resulting is a small sample variance value for that variable. The last part of the expression above, gives the formula that is most suitable for computation, either by hand or by a computer! Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 32 / 122
  • 33. Multivariate data analysis basics Parameters and descriptive statistics Dispersion: variance and standard deviation Since the sample variance is a function of the random data, the sample variance itself is a random quantity, and so has a population mean. In fact, the population mean of the sample variance is equal to the population variance: E(s2 j ) = σ2 j That is, the sample variance s2 j is unbiased for the population variance σ2 j . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 33 / 122
  • 34. Multivariate data analysis basics Parameters and descriptive statistics Dispersion: variance and standard deviation Some textbooks use a sample variance formula derived using maximum likelihood estimation principles. In this formula, the division is by n rather than n − 1. s2 j = n i=1(xij − ¯xj )2 n Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 34 / 122
  • 35. Multivariate data analysis basics Parameters and descriptive statistics Example Suppose that we have observed the following n = 5 resting pulse rates: 64, 68, 74, 76, 78. The sample mean is ¯x = 64+68+74+76+78 5 = 72. The maximum likelihood estimate of the variance is s2 = (64 − 72)2 + (68 − 72)2 + (74 − 72)2 + 5 +(76 − 72)2 + (78 − 72)2 5 = 136 5 = 27.2 The standard deviation based in this method is s = √ 27.2 = 5.215. The more commonly used variance estimate, the one given by statistical software, would be 136 5−1 = 34. The standard deviation would be s = √ 34 = 5.83. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 35 / 122
  • 36. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Association is concerned with how each variable is related to the other variable(s). In this case the first measure that we will consider is the covariance between two variables j and k. The population covariance is a measure of the association between pairs of variables in a population. Here, the population covariance between variables j and k is σjk = E{(Xij − µj )(Xik − µk )} Note that the product of the residuals (Xij − µj ) and (Xik − µk ) for variables j and k, respectively, is a function of the random variables Xij and Xik . Therefore, (Xij − µj )(Xik − µk ) is itself random, and has a population mean. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 36 / 122
  • 37. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation The population covariance is defined to be the population mean of this product of residuals. We see that if either both variables are greater than their respective means, or if they are both less than their respective means, then the product of the residuals will be positive. Thus, if the value of variable j tends to be greater than its mean when the value of variable k is larger than its mean, and if the value of variable j tends to be less than its mean when the value of variable k is smaller than its mean, then the covariance will be positive. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 37 / 122
  • 38. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Positive population covariances mean that the two variables are positively associated; variable j tends to increase with increasing values of variable k and vice-versa. Negative association can also occur. If one variable tends to be greater than its mean when the other variable is less than its mean, the product of the residuals will be negative, and you will obtain a negative population covariance. Variable j will tend to decrease with increasing values of variable k. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 38 / 122
  • 39. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation The population covariance σjk between variables j and k can be estimated by the sample covariance. This can be calculated using the formula below: sjk = 1 n − 1 n i=1 (Xij − ¯xj )(Xik − ¯xk ) = 1 n − 1 n i=1 Xij Xik − n n − 1 ¯x2 j ¯x2 k Just like in the formula for variance we have two expressions that make up this formula. The first half of the formula is most suitable for understanding the interpretation of the sample covariance, and the second half of the formula is what is used for calculation. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 39 / 122
  • 40. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Looking at the first half of the expression what you see inside the sum is the product of the residual differences between variable j and its mean times the residual differences between variable k and its mean. We can see that if either both variables tend to be greater than their respective means or less than their respective means, then the product of the residuals will tend to be positive leading to a positive sample covariance. Conversely if one variable takes values that are greater than its mean when the opposite variable takes a value less than its mean, then the product will take a negative value. In the end, when you add up this product over all of the observations, you will end up with a negative covariance. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 40 / 122
  • 41. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation So, in effect, a positive covariance would indicate a positive association between the variables j and k. And a negative association is when the covariance is negative. For computational purposes we will use the second half of the formula. For each subject, the product of the two variables is obtained, and then the products are summed to obtain the first term in the numerator. The second term in the numerator is obtained by taking the product of the sums of variable over the n subjects, then dividing the results by the sample size n. The difference between the first and second terms is then divided by n − 1 to obtain the covariance value. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 41 / 122
  • 42. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Again, sample covariance is a function of the random data, and hence, is random itself. As before, the population mean of the sample covariance sjk is equal the population covariance σjk ; i.e., E(sjk ) = σjk That is, the sample covariance sjk is unbiased for the population covariance σjk . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 42 / 122
  • 43. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation The sample covariance is a measure of the association between a pair of variables: sjk = 0 implies that the two variables are uncorrelated. (Note that this does not necessarily imply independence, we’ll get back to this later.) sjk > 0 implies that the two variables are positively correlated; i.e., values of variable j tend to increase with increasing values of variable k. The larger the covariance, the stronger the positive association between the two variables. sjk < 0 implies that the two variables are negatively correlated; i.e., values of variable j tend to decrease with increasing values of variable k. The smaller the covariance, the stronger the negative association between the two variables. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 43 / 122
  • 44. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Recall, that we had collected all of the population means of the p variables into a mean vector. Likewise, the population variances and covariances can be collected into the population variance-covariance matrix. This is also known by the name of population dispersion matrix. Σ =      σ2 1 σ12 . . . σ1p σ21 σ2 2 . . . σ2p ... ... ... ... σp1 σp2 . . . σ2 p      Note that the population variances appear along the diagonal of this matrix, and the covariance appear in the off-diagonal elements. So, the covariance between variables j and k will appear in row j and column k of this matrix. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 44 / 122
  • 45. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation The population variance-covariance matrix may be estimated by the sample variance-covariance matrix. The population variances and covariances in the above population variance-covariance matrix are replaced by the corresponding sample variances and covariances to obtain the sample variance-covariance matrix: S =      s2 1 s12 . . . s1p s21 s2 2 . . . s2p ... ... ... ... sp1 sp2 . . . s2 p      Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 45 / 122
  • 46. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Note that the sample variances appear along diagonal of this matrix and the covariances appear in the off-diagonal elements. So the covariance between variables j and k will appear in the jk-th element of this matrix. Notes: S (the sample variance-covariance matrix) is symmetric; i.e., sjk = skj . S is unbiased for the population variance covariance matrix Σ ; i.e., E(S) =      E(s2 1 ) E(s12) . . . E(s1p) E(s21) E(s2 2 ) . . . E(s2p) ... ... ... ... E(sp1) E(sp2) . . . E(s2 p )      = Σ Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 46 / 122
  • 47. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Since this matrix is a function of our random data, this means that the elements of this matrix are also going to be random, and the matrix on the whole is random as well. The statement “σ2 is unbiased”means that the mean of each element of that matrix is equal to the corresponding elements of the population. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 47 / 122
  • 48. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation In matrix notation sample variance-covariance matrix may be computed used the following expressions: S = 1 n − 1 n i=1 (Xi − ¯x)(Xi − ¯x) = 1 n − 1 n i=1 Xi Xi − n n − 1 ¯x¯x Just as we have seen in the previous formulas, the first half of the formula is used in interpretation, and the second half of the formula is what is used for calculation purposes. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 48 / 122
  • 49. Multivariate data analysis basics Parameters and descriptive statistics Measures of association: covariance and correlation Looking at the second term you can see that the first term in the numerator involves taking the data vector for each subject and multiplying by its transpose. The resulting matrices are then added over the n subjects. To obtain the second term in numerator, first compute the sum of the data vectors over the n subjects, then take the resulting vector and multiply by its transpose; then divide the resulting matrix by the squared number of subjects n2 . Take the difference between the two terms in the numerator and divide by n − 1. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 49 / 122
  • 50. Multivariate data analysis basics Parameters and descriptive statistics Example Suppose that we have observed the following n = 4 observations for variables x1 and x2. x1 x2 6 3 10 4 12 7 12 6 The sample means are x1 = 10 and x2 = 5. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 50 / 122
  • 51. Multivariate data analysis basics Parameters and descriptive statistics Example The maximum likelihood estimate of the covariance is the average product of deviations from the mean: s12 = (6 − 10)(3 − 5) + (10 − 10)(4 − 5)+ 4 +(12 − 10)(7 − 5) + (12 − 10)(6 − 5) 4 = 8 + 0 + 4 + 2 4 = 3.5 The positive value reflects the fact that as x1 increases, x2 also tends to increase. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 51 / 122
  • 52. Multivariate data analysis basics Parameters and descriptive statistics Problem Problem: The magnitude of the covariance value is not particularly helpful as it is a function of the magnitudes (scales) of the two variables. It is hard to distinguish the effects of the association between the two variables from the effects of their dispersions. Note, however, that the covariance between variables j and k must lie between the product of the two component standard deviations of variables j and k, and negative of that same product: −sj sk ≤ sjk ≤ sj sk . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 52 / 122
  • 53. Multivariate data analysis basics Parameters and descriptive statistics Example In an undergraduate statistics class, n = 30 females reported their heights (inches), and also measured their left forearm length (cm), left foot length (cm), and head circumference (cm). The sample variance-covariance matrix is the following: Notice that the matrix has four rows and four columns because there are four variables being considered. Also notice that the matrix is symmetric. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 53 / 122
  • 54. Multivariate data analysis basics Parameters and descriptive statistics Example Here are a few examples of the information in the matrix: The variance of the height variable is 8.74. Thus the standard deviation is√ 8.74 = 2.956. The variance of the left foot measurement is 1.908 (in the 3rd diagonal element). Thus the standard deviation for this variable is √ 1.908 = 1.381. The covariance between height and left arm is 3.022, found in the 1st row, 2nd column and also in the 2nd row, 1st column. The covariance between left foot and left arm is 1.234, found in the 3rd row, 2nd column and also in the 2nd row, 3rd column. All covariance values are positive so all pairwise associations are positive. But, the magnitudes do not tell us about the strength of the associations. To assess the strength of an association, we use correlation values. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 54 / 122
  • 55. Multivariate data analysis basics Parameters and descriptive statistics Correlation This suggests an alternative measure of association. The population correlation is defined to be equal to the population covariance divided by the product of the population standard deviations: ρjk = σjk σj σk . The population correlation may be estimated by substituting into the formula the sample covariances and standard deviations: rjk = sij si sj = n i=1 Xij Xik − ( n i=1 Xij )( n i=1 Xik )/n2 ( n i=1 X2 ij − ( n i=1 Xij /n)2)( n i=1 X2 ik − ( n i=1 Xik /n)2) Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 55 / 122
  • 56. Multivariate data analysis basics Parameters and descriptive statistics Correlation It is very important to note that the population as well as the sample correlation must lie between -1 and 1. −1 ≤ ρjk ≤ 1 −1 ≤ rjk ≤ 1 Therefore: ρjk = 0 indicates, as you might expect, that the two variables are uncorrelated . ρjk close to +1 will indicate a strong positive dependence ρjk close to -1 indicates a strong negative dependence. Sample correlation coefficients also have similar interpretation. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 56 / 122
  • 57. Multivariate data analysis basics Parameters and descriptive statistics Correlation matrix For a collection of p variables, the correlation matrix is a p × p matrix that displays the correlations between pairs of variables. For instance, the value in the j-th row and k-th column gives the correlation between variables xj and xk . The correlation matrix is symmetric so that the value in the k-th row and j-th column is also the correlation between variables xj and xk . The diagonal elements of the correlation matrix are all identically equal to 1. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 57 / 122
  • 58. Multivariate data analysis basics Parameters and descriptive statistics Correlation matrix The sample correlation matrix is denoted as R. R =      1 r12 . . . r1p r21 1 . . . r2p ... ... ... ... rp1 rp2 . . . 1      Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 58 / 122
  • 59. Multivariate data analysis basics Parameters and descriptive statistics Example The following covariance matrix show the pairwise covariances for the height, left forearm, left foot and head circumference measurements of n = 30 female college students. Here are two examples of calculating a correlation coefficient: The correlation between height and left forearm is 3.022√ 8.74 √ 2.402 = 0.66. The correlation between head circumference and left foot is 0.118√ 3.434 √ 1.908 = 0.046. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 59 / 122
  • 60. Multivariate data analysis basics Parameters and descriptive statistics Example The complete sample correlation matrix for this example is the following: Overall, we see moderately strong linear associations among the variables height, left arm and left foot and quite weak (almost 0) associations between head circumference and the other three variables. In practice, use scatter plots of the variables to fully understand the associations between variables. It is not a good idea to rely on correlations without seeing the plots. Correlation values are affected by outliers and curvilinearity. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 60 / 122
  • 61. Multivariate data analysis basics Parameters and descriptive statistics Overall measures of dispersion Sometimes it is also useful to have an overall measure of dispersion in the data. In this measure it would be good to include all of the variables simultaneously, rather than one at a time. In the past we looked at the individual variables and their variances to measure the individual variances. Here we are going to look at measures of dispersion of all variables together, particularly we are going to look at such measures that look at total variation. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 61 / 122
  • 62. Multivariate data analysis basics Parameters and descriptive statistics Overall measures of dispersion The variance σ2 j measures the dispersion of an individual variable Xj . The following two are used to measure the dispersion of all variables together. Total Variation Generalized Variance To understand total variation we first must find the trace of a square matrix. A square matrix is a matrix that has an equal number of columns and rows. Important examples of square matrices include the variance-covariance and correlation matrices. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 62 / 122
  • 63. Multivariate data analysis basics Parameters and descriptive statistics Overall measures of dispersion The trace of an n × n matrix A is trace(A) = n i=1 aii . For instance, in a 10 × 10 matrix, the trace is the sum of the diagonal elements. The total variation, therefore, of a random vector X is simply the trace of the population variance-covariance matrix. trace(Σ) = σ2 1 + σ2 2 + . . . σ2 p. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 63 / 122
  • 64. Multivariate data analysis basics Parameters and descriptive statistics Overall measures of dispersion Thus, the total variation is equal to the sum of the population variances. The total variation can be estimated by: trace(S) = s2 1 + s2 2 + · · · + s2 p . The total variation is of interest for principal components analysis and factor analysis and we will look at these concepts later in this course. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 64 / 122
  • 65. Multivariate data analysis basics Parameters and descriptive statistics Example Let us use the data from the USDA women’s health survey again to illustrate this. We have taken the variances for each of the variables from software output and have placed them in the table below. Variable Variance Calcium 157,829.4 Iron 35.8 Protein 934.9 Vitamin A 2,668,452.4 Vitamin C 5,416.3 Total 2,832,668.8 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 65 / 122
  • 66. Multivariate data analysis basics Parameters and descriptive statistics Example The total variation for the nutrient intake data is determined by simply adding up all of the variances for each of the individual variables. The total variation equals 2,832,668.8. This is a very large number. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 66 / 122
  • 67. Multivariate data analysis basics Parameters and descriptive statistics Example Problem: The problem with total variation is that it does not take into account correlations among the variables. These plots show simulated data for pairs of variables with different levels of correlation. In each case, the variances for both variables are equal to 1, so that the total variation is 2. The corresponding variance-covariance matrix is: Σ = 1 ρjk σj σk ρjk σj σk 1 where ρjk is the correaltion coefficient between the two variables. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 67 / 122
  • 68. Multivariate data analysis basics Parameters and descriptive statistics Example When the correlation r = 0, then we see a shotgun-blast pattern of points, widely dispersed over the entire range of the plot. Σ = 1 0 0 1 , being the Total Variation trace(Σ) = 2 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 68 / 122
  • 69. Multivariate data analysis basics Parameters and descriptive statistics Example Increasing the correlation to r = 0.7, we see an oval-shaped pattern. Note that the points are not as widely dispersed. Σ = 1 0.7 0.7 1 , being the Total Variation trace(Σ) = 2 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 69 / 122
  • 70. Multivariate data analysis basics Parameters and descriptive statistics Example Increasing the correlation to r = 0.9, we see that the points fall along a 45 degree line, and are even less dispersed. Σ = 1 0.9 0.9 1 , being the Total Variation trace(Σ) = 2 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 70 / 122
  • 71. Multivariate data analysis basics Parameters and descriptive statistics Generalized variance To take into account the correlations among pairs of variables an alternative measure of overall variance is suggested. This measure takes a large value when the various variables show very little correlation among themselves. In contrast, this measure takes a small value if the variables show very strong correlation among themselves, either positive or negative. This particular measure of dispersion is the generalized variance. In order to define the generalized variance we first define the determinant of the matrix. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 71 / 122
  • 72. Multivariate data analysis basics Parameters and descriptive statistics Generalized variance We will start simple with a 2 × 2 matrix and then we will move on to more general definitions for larger matrices. Let us consider the determinant of a 2 × 2 matrix B as shown below. Here we can see that it is the product of the two diagonal elements minus the product of the off-diagonal elements. |B| = b11 b12 b21 b22 = b11b22 − b12b21. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 72 / 122
  • 73. Multivariate data analysis basics Parameters and descriptive statistics Generalized variance Here is an example of a simple matrix that has the elements 5, 1, 2 and 4. You will get the determinant 18. The product of the diagonal 5 × 4 subtracting the elements of the off-diagonal 1 × 2 yields an answer of 18: 5 1 2 4 = 5 × 4 − 1 × 2 = 20 − 2 = 18. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 73 / 122
  • 74. Multivariate data analysis basics Parameters and descriptive statistics Generalized variance More generally the determinant of a general p × p matrix B is given by the expression shown below: |B| = p j=1 (−1)j+1 b1j |B1j |. The expression involves the sum over all of the first row of B. Note that these elements are noted by b1j . These are pre-multiplied by −1 raised to the (j + 1)-th power, so that basically we are going to have alternating plus and minus signs in our sum. The matrix B1j is obtained by deleting row i and column j from the matrix B. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 74 / 122
  • 75. Multivariate data analysis basics Parameters and descriptive statistics Generalized variance For a 3 × 3 matrix the determinant is: |B| = b11 b12 b13 b21 b22 b23 b31 b32 b33 = b11b22b33 + b12b23b31 + b13b21b32 −b13b22b31 − b32b23b11 − b33b21b12. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 75 / 122
  • 76. Multivariate data analysis basics Parameters and descriptive statistics Generalized variance By definition the generalized variance of a random vector X is equal to |Σ|, the determinant of the variance-covariance matrix. The generalized variance can be estimated by calculating |S|, the determinant of the sample variance-covariance matrix. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 76 / 122
  • 77. Multivariate data analysis basics Matrix manipulations for sample statistics Subsection 3 Matrix manipulations for sample statistics Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 77 / 122
  • 78. Multivariate data analysis basics Matrix manipulations for sample statistics The Mean and mean corrected data Suppose we have a n × p data matrix X and a 1 × n vector of ones, 1 . The mean vector is given by: ¯x = 1 n 1 X, and the mean corrected data are given by Xm = X − 1¯x where Xm gives the matrix containing the mean-corrected data. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 78 / 122
  • 79. Multivariate data analysis basics Matrix manipulations for sample statistics Sum of squares of cross products and covariance The matrix of sum of squares and cross products (SSCP) SSCPm is given by SSCPm = XmXm and the sample variance-covariance matrix is given by S = 1 n − 1 SSCPm = 1 n − 1 XmXm Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 79 / 122
  • 80. Multivariate data analysis basics Matrix manipulations for sample statistics Standardized data If we define a p × p diagonal matrix D, which has variances of the variables in the diagonal, then standardized data are given by Xs = XmD−1/2 . The SSCPs of standardized data is given by SSCPs = XsXs Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 80 / 122
  • 81. Multivariate data analysis basics Matrix manipulations for sample statistics Correlation matrix The correlation matrix is given by R = 1 n − 1 SSCPs = 1 n − 1 XsXs The generalized variance is given by |S| All these computations can be easily performed in R using the matrix operations... The following table presents and hypothetical data set Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 81 / 122
  • 82. Multivariate data analysis basics Matrix manipulations for sample statistics Firm Original Mean-corrected Standardized data data data X1 X2 X1 X2 X1 X2 1 13.000 4.000 7.917 3.833 1.619 1.108 2 10.000 6.000 4.917 5.833 1.006 1.686 3 10.000 2.000 4.917 1.833 1.006 0.530 4 8.000 -2.000 2.917 -2.167 0.597 -0.626 5 7.000 4.000 1.917 3.833 0.392 1.108 6 6.000 -3.000 0.917 -3.167 0.187 -0.915 7 5.000 0.000 -0.083 -0.167 -0.017 -0.048 8 4.000 2.000 -1.083 1.833 -0.222 0.530 9 2.000 -1.000 -3.083 -1.167 -0.631 -0.337 10 0.000 -5.000 -5.083 -5.167 -1.040 -1.493 11 -1.000 -1.000 -6.083 -1.167 -1.244 -0.337 12 -3.000 -4.000 -8.083 -4.167 -1.653 -1.204 Mean 5.083 0.167 0.000 0.000 0.000 0.000 SS 262.917 131.667 11.000 11.000 Var 23.902 11.970 23.902 11.970 1.000 1.000 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 82 / 122
  • 83. Multivariate data analysis basics Matrix manipulations for sample statistics Example The SSCP matrix of this data set is: SSCP = 262.917 136.375 136.375 131.667 The S matrix is: S = 1 n − 1 SSCP = 23.902 12.398 12.398 11.970 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 83 / 122
  • 84. Multivariate data analysis basics Matrix manipulations for sample statistics Example The correlation matrix R of these data is: R = s2 1 s1s1 s12 s1s2 s21 s2s1 s2 2 s2s2 = 1 0.733 0.733 1 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 84 / 122
  • 85. Multivariate data analysis basics A review of matrix definitions and operations Subsection 4 A review of matrix definitions and operations Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 85 / 122
  • 86. Multivariate data analysis basics A review of matrix definitions and operations Just a review! The following material reviews basic matrix definitions and operations. You should have previous knowledge on matrix algebra. This review does not relieve you from responsibility of being sufficiently proficient in these matters! Therefore be cautious and if you do not feel comfortable with these issues please get back again to these matters! Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 86 / 122
  • 87. Multivariate data analysis basics A review of matrix definitions and operations Transpose of a matrix The transpose of a matrix A is a matrix in which the rows of the transpose are the columns of A (so the columns of the transpose are the rows of A). The transpose of A is denoted as A . Example: for the 4 × 2 matrix, A =     3 6 2 5 3 7 1 8     the transpose is the 2 × 4 matrix A = 3 2 3 1 6 5 7 8 Notice that the columns of A are the rows of A Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 87 / 122
  • 88. Multivariate data analysis basics A review of matrix definitions and operations Symmetric matrices A square matrix A is symmetric if aij = aji , for all i and j. That is throughout the matrix, the element in the i-th row and j-th column equals the element in the j-th row and i-th column. Said another way, a matrix is symmetric when, A = A. Important examples of symmetric matrices in multivariate statistics include the variance-covariance matrix and the correlation matrix. These were defined when we considered descriptive statistics. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 88 / 122
  • 89. Multivariate data analysis basics A review of matrix definitions and operations Symmetric matrices Example: the following is an example of a symmetric matrix. Notice that the first row and the first column are identical, the second row and the second column are the same, and that the third row is the same as the third column.   4 2 1 2 3 0 1 0 5   Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 89 / 122
  • 90. Multivariate data analysis basics A review of matrix definitions and operations Adding two matrices Two matrices may be added if and only if they have the same dimensions (same number of rows and also same number of columns as each other. To add two matrices, add corresponding elements (in terms of location). Example: 2 4 6 8 + 1 3 5 7 = 2 + 1 4 + 3 6 + 5 8 + 7 = 3 7 11 15 Note: one matrix is subtracted from another in the same way that matrices are added. To do the subtraction, subtract each element of the first matrix from the corresponding (in location) element of the first matrix Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 90 / 122
  • 91. Multivariate data analysis basics A review of matrix definitions and operations Multiplying a matrix by a scalar The word scalar is a synonym for a numerical constant. In matrix terms, a scalar is a matrix with one row and one column. To multiply a matrix by a scalar, multiply each element in the matrix by the scalar. Example: in this example, we multiply a matrix by the value 3. 3 × 2 4 6 8 = 3 × 2 3 × 4 3 × 6 3 × 8 = 6 12 18 24 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 91 / 122
  • 92. Multivariate data analysis basics A review of matrix definitions and operations Multiplication of matrices To be able to perform the matrix multiplication C = A × B, the number of columns in the matrix A must equal the number of rows in the matrix B. The element in the i-th row and j-th column of the answer matrix C, is the cross product sum of the i-th row of the matrix A and the j-th column of the matrix B. This is done of all combinations of rows of A and columns of B. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 92 / 122
  • 93. Multivariate data analysis basics A review of matrix definitions and operations Multiplication of matrices Example: we’ll carry out the multiplication A×B =   1 2 2 3 4 3  × 4 2 1 3 . The answer is C =   1 × 4 + 2 × 1 1 × 2 + 2 × 3 2 × 4 + 3 × 1 2 × 2 + 3 × 3 4 × 4 + 3 × 1 4 × 2 + 3 × 3   =   6 8 11 13 19 17   . Notice, for instance, that the element in the 1st row and 1st column of the answer C is the cross product sum of the 1st row of A and the 1st row of B. As another example, the 2nd row, 1st column element of C is the cross product sum of the 2nd row of A and the 1st column of B. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 93 / 122
  • 94. Multivariate data analysis basics A review of matrix definitions and operations The Identity Matrix An identity matrix is a square matrix that has the value one in each main diagonal position (from upper left to bottom right) and has the value 0 in all other locations. As an example of an identity matrix, the 3 × 3 identity matrix is I =   1 0 0 0 1 0 0 0 1   . I is called the identity matrix because multiplication of any square matrix A by the identity matrix yields the original matrix A as the answer. That is AI = IA = A. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 94 / 122
  • 95. Multivariate data analysis basics A review of matrix definitions and operations Matrix inverse An inverse (in the traditional sense) can be found only for square matrices. The inverse of a square matrix A is the matrix A−1 such that A−1 A = AA−1 = I. The calculation of an inverse for large matrices is a laborious process that we’ll leave to the computer. For 2 × 2 matrices, however, the formula is relatively simple. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 95 / 122
  • 96. Multivariate data analysis basics A review of matrix definitions and operations Matrix inverse For A = a11 a12 a21 a22 , the inverse is A−1 = 1 a11a22 − a12a21 a22 −a12 −a21 a11 . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 96 / 122
  • 97. Multivariate data analysis basics A review of matrix definitions and operations Matrix inverse Example: we’ll determine the inverse of A = 10 6 8 5 . The inverse is A−1 = 1 10 × 5 − 6 × 8 5 −6 −8 10 = 1 2 5 −6 −8 10 = 2.5 −3 −4 5 . You might want to check that A−1 A = I. (It does!) Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 97 / 122
  • 98. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition Let A be a p × p square matrix and I be the p × p identity matrix. The scalars λ1, λ2, ..., λp satisfying the characteristic equation |A − λI| = 0 are called the eigenvalues (or characteristic roots) of matrix A. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 98 / 122
  • 99. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition On the left-hand side, we have the matrix A minus λ times the Identity matrix. When we calculate the determinant of the resulting matrix, we end up with a polynomial of order p. Setting this polynomial equal to zero, and solving for λ we obtain the desired eigenvalues. In general, we will have p solutions and so there are p eigenvalues, not necessarily all unique. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 99 / 122
  • 100. Multivariate data analysis basics A review of matrix definitions and operations Example Let A = 1 0 1 3 . Then |A − λI| = 1 0 1 3 − λ 1 0 0 1 = 1 − λ 0 1 3 − λ = (1 − λ)(3 − λ) implies that there are two roots, λ1 = 1 and λ2 = 3. The eigenvalues of A are 3 and 1 (the eigenvalues are in generally listed in descending order). Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 100 / 122
  • 101. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition The corresponding eigenvectors e1, e2, ..., ep are obtained by solving the expression below: (A − λj I)ej = 0 Here, we have the difference between the matrix A minus the j-th eigenvalue times the Identity matrix, this quantity is then multiplied by the j-th eigenvector and set it all equal to zero. This will obtain the eigenvector ej associated with eigenvalue λj . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 101 / 122
  • 102. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition Note: This does not have a generally have a unique solution. So, to obtain a unique solution we will often require that ej transposed ej is equal to 1. Or, if you like, the sum of the square elements of ej is equal to 1, which means the length (norm) of the eigenvector is the unity. ej ej = 1 Also note that eigenvectors corresponding to different eigenvalues are orthogonal. In situations, where two (or more) eigenvalues are equal, corresponding eigenvectors will still be orthogonal. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 102 / 122
  • 103. Multivariate data analysis basics A review of matrix definitions and operations Example (continued) The eigenvectors of matrix A associated to those eigenvalues (1 and 3) can be determined by solving the following equations: Ax = λ1x 1 0 1 3 x1 x2 = 1 × x1 x2 x1 x1 + 3x2 = x1 x2 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 103 / 122
  • 104. Multivariate data analysis basics A review of matrix definitions and operations Example (continued) There are many solutions! Therefore, setting x2 = 1 (arbitrarily) gives x1 = −2, and hence −2 1 is an eigenvector corresponding to the eigenvalue 1. After transforming for unit norm: e = (−2/ √ 5 1/ √ 5). Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 104 / 122
  • 105. Multivariate data analysis basics A review of matrix definitions and operations Example (continued) Ax = λ2x 1 0 1 3 x1 x2 = 3 × x1 x2 x1 x1 + 3x2 = 3x1 3x2 Again, setting x2 = 1 and x1 = 0 0 1 is an eigenvector corresponding to the eigenvalue 3. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 105 / 122
  • 106. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition Any symmetric square matrix can be reconstructed from its eigenvalues and eigenvectors. Let A be a p × p symmetric matrix. Then A can be expressed in terms of its p eigenvalue-eigenvector pairs (λi , ei ) as A = p j=1 λj ej ej Or, if you dispose the eigenvalues in a diagonal matrix Λ and the eigenvectors in a matrix E, then A = EΛE . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 106 / 122
  • 107. Multivariate data analysis basics A review of matrix definitions and operations Example For example, let A = 2.2 0.4 0.4 2.8 Then |A − λI| = λ2 − 5λ + 6.16 − 0.16 = (λ − 3)(λ − 2) so A has eigenvalues λ1 = 3 and λ2 = 2. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 107 / 122
  • 108. Multivariate data analysis basics A review of matrix definitions and operations Example The corresponding eigenvectors are e1 = ( 1/ √ 5 2/ √ 5 ) and e2 = ( −2/ √ 5 1/ √ 5 ), respectively. Therefore A = 2.2 0.4 0.4 2.8 = 3 1/ √ 5 2/ √ 5 1/ √ 5 2/ √ 5 + +2 −2/ √ 5 1/ √ 5 −2/ √ 5 1/ √ 5 = 0.6 1.2 1.2 2.4 + 1.6 −0.8 −0.8 0.4 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 108 / 122
  • 109. Multivariate data analysis basics A review of matrix definitions and operations Example Or, A = 2.2 0.4 0.4 2.8 = = 1/ √ 5 −2/ √ 5 2/ √ 5 1/ √ 5 3 0 0 2 1/ √ 5 2/ √ 5 −2/ √ 5 1/ √ 5 = 3/ √ 5 −4/ √ 5 6/ √ 5 2/ √ 5 1/ √ 5 −2/ √ 5 2/ √ 5 1/ √ 5 = 3/5 + 8/5 6/5 − 4/5 6/5 − 4/5 12/5 + 2/5/ Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 109 / 122
  • 110. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition Usually the matrix A is taken to be either the variance-covariance matrix Σ, or the correlation matrix, or their estimates S and R, respectively. Eigenvalues and eigenvectors are used for: Computing prediction and confidence ellipses Principal Components Analysis (later in the course) Factor Analysis (also later in this course) Discriminant analysis (also later in this course) Multidimensional scaling (also later in this course) ... Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 110 / 122
  • 111. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Consider the correlation matrix R as shown below: R = 1 ρ ρ 1 . Then, using the definition of the eigenvalues, we must calculate the determinant of R − λ times the Identity matrix. |R − λI| = 1 ρ ρ 1 − λ 1 0 0 1 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 111 / 122
  • 112. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Carrying out the math we end up with the matrix with 1 − λ on the diagonal and ρ on the off-diagonal. Then calculating this determinant we obtain 1 − λ squared minus ρ2 : 1 − λ ρ ρ 1 − λ = (1 − λ)2 − ρ2 = λ2 − 2λ + 1 − ρ2 . Setting this expression equal to zero we end up with the following... λ2 − 2λ + 1 − ρ2 = 0. Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 112 / 122
  • 113. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case To solve for λ we use the general result that any solution to the second order polynomial below: ay2 + by + c = 0 is given by the following expression: y = −b ± √ b2 − 4ac 2a . Here, a = 1, b = −2 (the term that precedes λ and c is equal to 1 − ρ2 . Substituting these terms in the equation above, we obtain that λ must be equal to 1 plus or minus the correlation ρ. λ = 2 ± 22 − 4(1 − ρ2) 2 = 1 ± 1 − (1 − ρ2) = 1 ± ρ Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 113 / 122
  • 114. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case We will take the following solutions: λ1 = 1 + ρ λ2 = 1 − ρ Next, to obtain the corresponding eigenvectors, we must solve a system of equations below: (R − λ1I)e1 = 0 and (R − λ2I)e2 = 0 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 114 / 122
  • 115. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Or in other words, this is translated for this specific problem in the expression below: 1 ρ ρ 1 − λ1 1 0 0 1 e11 e12 = 0 0 This simplifies as follows: 1 − λ1 ρ ρ 1 − λ1 e11 e12 = 0 0 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 115 / 122
  • 116. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Yielding a system of two equations with two unknowns: (1 − λ1)e11 + ρe12 = 0 ρe11 + (1 − λ1)e12 = 0 Note that this does not have a unique solution. If (e11, e12) is one solution, then a second solution can be obtained by multiplying the first solution by any non-zero constant c, i.e., (ce11, ce12). Therefore, we will require the additional condition that the sum of the squared values of e11 and e12 are equal to 1. e2 11 + e2 12 = 1 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 116 / 122
  • 117. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Consider the first equation: (1 − λ1)e11 + ρe12 = 0 Solving this equation for e12 and we obtain the following: e12 = − (1 − λ1) ρ e11 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 117 / 122
  • 118. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Substituting this into e2 11 + e2 12 = 1 we get the following: e2 11 + (1 − λ1)2 ρ2 e2 11 = 1 Recall that λ = 1 ± ρ. In either case we end up finding that (1 − λ1)2 = ρ2 , so that the expression above simplifies to: 2e2 11 = 1 Or, in other words: e11 = 1 √ 2 Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 118 / 122
  • 119. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Using the expression for e12 which we obtained above, e12 = − 1 − λ1 ρ e11 = 1 √ 2 we get e11 = 1 √ 2 and e12 = 1 √ 2 for λ1 = 1 + ρ Using the same reasoning we get e21 = 1 √ 2 and e22 = − 1 √ 2 for λ2 = 1 − ρ Therefore, the two eigenvectors are given by the two vectors as shown below: 1√ 2 1√ 2 for λ = 1 + ρ and 1√ 2 − 1√ 2 for λ = 1 − ρ Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 119 / 122
  • 120. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Some properties of the eigenvalues of the variance-covariance matrix are to be considered at this point. Suppose that λ1, λ2, ..., λp are the eigenvalues of the variance-covariance matrix Σ. By definition, the total variation is given by the sum of the variances. It turns out that this is also equal to the sum of the eigenvalues of the variance-covariance matrix. Thus, the total variation is: p j=1 s2 j = s2 1 + s2 2 + · · · + s2 p = λ1 + λ2 + · · · + λp = p j=1 λj Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 120 / 122
  • 121. Multivariate data analysis basics A review of matrix definitions and operations Eigenvalues, eigenvectors and spectral decomposition: a special case Some properties of the eigenvalues of the variance-covariance matrix are to be considered at this point. Suppose that λ1, λ2, ..., λp are the eigenvalues of the variance-covariance matrix Σ. The generalized variance is equal to the product of the eigenvalues: |Σ| = p j=1 λj = λ1 × λ2 × · · · × λp Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 121 / 122
  • 122. Multivariate data analysis basics A review of matrix definitions and operations Quadratic forms A quadratic form Q(x) in the p variables x1, x2, ...xp is Q(x) = x Ax, where x = [x1, x2, ..., xp] and A is a p × p symmetric matrix. Note that a quadratic form can be written as Q(x) = p i=1 p j=1 aij xi xj . For example, Q(x) = x1 x2 1 1 1 1 x1 x2 = x2 1 + x1x2 + x2 2 . Q(x) = x1 x2 x3   1 3 0 3 −1 −2 0 −2 2     x1 x2 x3   = x2 1 + 6x1x2 − x2 2 − 4x2x3 + 2x2 3 . Jorge M. Mendes (NOVAIMS) Multivariate Data Analysis Basics 2015 122 / 122