SlideShare a Scribd company logo
1 of 30
Correlation and
Regression
Davina Bristow &
Angela Quayle
Topics Covered:
 Is there a relationship between x and y?
 What is the strength of this relationship
 Pearson’s r
 Can we describe this relationship and use this to predict y from
x?
 Regression
 Is the relationship we have described statistically significant?
 t test
 Relevance to SPM
 GLM
The relationship between x and y
 Correlation: is there a relationship between 2
variables?
 Regression: how well a certain independent
variable predict dependent variable?
 CORRELATION  CAUSATION
In order to infer causality: manipulate independent
variable and observe effect on dependent variable
Scattergrams
Y
X
Y
X
Y
X
Y
Y Y
Positive correlation Negative correlation No correlation
Variance vs Covariance
 First, a note on your sample:
 If you’re wishing to assume that your sample is
representative of the general population (RANDOM
EFFECTS MODEL), use the degrees of freedom (n – 1)
in your calculations of variance or covariance.
 But if you’re simply wanting to assess your current
sample (FIXED EFFECTS MODEL), substitute n for
the degrees of freedom.
Variance vs Covariance
 Do two variables change together?
1
)
)(
(
)
,
cov( 1






n
y
y
x
x
y
x
i
n
i
i
Covariance:
• Gives information on the degree to
which two variables vary together.
• Note how similar the covariance is to
variance: the equation simply
multiplies x’s error scores by y’s error
scores as opposed to squaring x’s error
scores.
1
)
( 2
1
2





n
x
x
S
n
i
i
x
Variance:
• Gives information on variability of a
single variable.
Covariance
 When X and Y : cov (x,y) = pos.
 When X and Y : cov (x,y) = neg.
 When no constant relationship: cov (x,y) = 0
1
)
)(
(
)
,
cov( 1






n
y
y
x
x
y
x
i
n
i
i
Example Covariance
x y x
xi
 y
yi
 ( x
i
x  )( y
i
y  )
0 3 -3 0 0
2 2 -1 -1 1
3 4 0 1 0
4 0 1 -3 -3
6 6 3 3 9
3

x 3

y  7
75
.
1
4
7
1
))
)(
(
)
,
cov( 1








n
y
y
x
x
y
x
i
n
i
i What does this
number tell us?
0
1
2
3
4
5
6
7
0 1 2 3 4 5 6 7
Problem with Covariance:
 The value obtained by covariance is dependent on the size of
the data’s standard deviations: if large, the value will be
greater than if small… even if the relationship between x and y
is exactly the same in the large versus small standard
deviation datasets.
Example of how covariance value
relies on variance
High variance data Low variance data
Subject x y x error * y
error
x y X error * y
error
1 101 100 2500 54 53 9
2 81 80 900 53 52 4
3 61 60 100 52 51 1
4 51 50 0 51 50 0
5 41 40 100 50 49 1
6 21 20 900 49 48 4
7 1 0 2500 48 47 9
Mean 51 50 51 50
Sum of x error * y error : 7000 Sum of x error * y error : 28
Covariance: 1166.67 Covariance: 4.67
Solution: Pearson’s r
 Covariance does not really tell us anything
 Solution: standardise this measure
 Pearson’s R: standardises the covariance value.
 Divides the covariance by the multiplied standard deviations of
X and Y:
y
x
xy
s
s
y
x
r
)
,
cov(

Pearson’s R continued
1
)
)(
(
)
,
cov( 1






n
y
y
x
x
y
x
i
n
i
i
y
x
i
n
i
i
xy
s
s
n
y
y
x
x
r
)
1
(
)
)(
(
1






1
*
1




n
Z
Z
r
n
i
y
x
xy
i
i
Limitations of r
 When r = 1 or r = -1:
 We can predict y from x with certainty
 all data points are on a straight line: y = ax + b
 r is actually
 r = true r of whole population
 = estimate of r based on data
 r is very sensitive to extreme values:
0
1
2
3
4
5
0 1 2 3 4 5 6
r̂
r̂
Regression
 Correlation tells you if there is an association
between x and y but it doesn’t describe the
relationship or allow you to predict one
variable from the other.
 To do this we need REGRESSION!
Best-fit Line
= ŷ, predicted value
 Aim of linear regression is to fit a straight line, ŷ = ax + b, to data that
gives best prediction of y for any value of x
 This will be the line that
minimises distance between
data and fitted line, i.e.
the residuals
intercept
ε
ŷ = ax + b
ε = residual error
= y i , true value
slope
Least Squares Regression
 To find the best line we must minimise the sum of
the squares of the residuals (the vertical distances
from the data points to our line)
Residual (ε) = y - ŷ
Sum of squares of residuals = Σ (y – ŷ)2
Model line: ŷ = ax + b
 we must find values of a and b that minimise
Σ (y – ŷ)2
a = slope, b = intercept
Finding b
 First we find the value of b that gives the min
sum of squares
ε ε
b
b
b
 Trying different values of b is equivalent to
shifting the line up and down the scatter plot
Finding a
 Now we find the value of a that gives the min
sum of squares
b b b
 Trying out different values of a is equivalent to
changing the slope of the line, while b stays
constant
Minimising sums of squares
 Need to minimise Σ(y–ŷ)2
 ŷ = ax + b
 so need to minimise:
Σ(y - ax - b)2
 If we plot the sums of squares
for all different values of a and b
we get a parabola, because it is a
squared term
 So the min sum of squares is at
the bottom of the curve, where
the gradient is zero.
Values of a and b
sums
of
squares
(S)
Gradient = 0
min S
The maths bit
 The min sum of squares is at the bottom of the curve
where the gradient = 0
 So we can find a and b that give min sum of squares
by taking partial derivatives of Σ(y - ax - b)2 with
respect to a and b separately
 Then we solve these for 0 to give us the values of a
and b that give the min sum of squares
The solution
 Doing this gives the following equations for a and b:
a =
r sy
sx
r = correlation coefficient of x and y
sy = standard deviation of y
sx = standard deviation of x
 From you can see that:
 A low correlation coefficient gives a flatter slope (small value of
a)
 Large spread of y, i.e. high standard deviation, results in a
steeper slope (high value of a)
 Large spread of x, i.e. high standard deviation, results in a flatter
slope (high value of a)
The solution cont.
 Our model equation is ŷ = ax + b
 This line must pass through the mean so:
y = ax + b b = y – ax
 We can put our equation for a into this giving:
b = y – ax
b = y -
r sy
sx
r = correlation coefficient of x and y
sy = standard deviation of y
sx = standard deviation of x
x
 The smaller the correlation, the closer the
intercept is to the mean of y
Back to the model
 If the correlation is zero, we will simply predict the mean of y for every
value of x, and our regression line is just a flat straight line crossing the
x-axis at y
 But this isn’t very useful.
 We can calculate the regression line for any data, but the important
question is how well does this line fit the data, or how good is it at
predicting y from x
ŷ = ax + b =
r sy
sx
r sy
sx
x + y - x
r sy
sx
ŷ = (x – x) + y
Rearranges to:
a b
a a
How good is our model?
 Total variance of y: sy
2 =
∑(y – y)2
n - 1
SSy
dfy
=
 Variance of predicted y values (ŷ):
 Error variance:
sŷ
2 =
∑(ŷ – y)2
n - 1
SSpred
dfŷ
=
This is the variance
explained by our
regression model
serror
2 =
∑(y – ŷ)2
n - 2
SSer
dfer
=
This is the variance of the error
between our predicted y values and
the actual y values, and thus is the
variance in y that is NOT explained
by the regression model
 Total variance = predicted variance + error variance
sy
2 = sŷ
2 + ser
2
 Conveniently, via some complicated rearranging
sŷ
2 = r2 sy
2
r2 = sŷ
2 / sy
2
 so r2 is the proportion of the variance in y that is explained by
our regression model
How good is our model cont.
How good is our model cont.
 Insert r2 sy
2 into sy
2 = sŷ
2 + ser
2 and rearrange to get:
ser
2 = sy
2 – r2sy
2
= sy
2 (1 – r2)
 From this we can see that the greater the correlation
the smaller the error variance, so the better our
prediction
Is the model significant?
 i.e. do we get a significantly better prediction of y
from our regression equation than by just predicting
the mean?
 F-statistic:
F(dfŷ,dfer) =
sŷ
2
ser
2
=......=
r2 (n - 2)2
1 – r2
complicated
rearranging
 And it follows that:
t(n-2) =
r (n - 2)
√1 – r2
(because F = t2)
So all we need to
know are r and n
General Linear Model
 Linear regression is actually a form of the
General Linear Model where the parameters
are a, the slope of the line, and b, the intercept.
y = ax + b +ε
 A General Linear Model is just any model that
describes the data in terms of a straight line
Multiple regression
 Multiple regression is used to determine the effect of a number
of independent variables, x1, x2, x3 etc, on a single dependent
variable, y
 The different x variables are combined in a linear way and
each has its own regression coefficient:
y = a1x1+ a2x2 +…..+ anxn + b + ε
 The a parameters reflect the independent contribution of each
independent variable, x, to the value of the dependent variable,
y.
 i.e. the amount of variance in y that is accounted for by each x
variable after all the other x variables have been accounted for
SPM
 Linear regression is a GLM that models the effect of one
independent variable, x, on ONE dependent variable, y
 Multiple Regression models the effect of several independent
variables, x1, x2 etc, on ONE dependent variable, y
 Both are types of General Linear Model
 GLM can also allow you to analyse the effects of several
independent x variables on several dependent variables, y1, y2,
y3 etc, in a linear combination
 This is what SPM does and all will be explained next week!

More Related Content

Similar to Cr-and-Regress.ppt

regression and correlation
regression and correlationregression and correlation
regression and correlationPriya Sharma
 
REGRESSION ANALYSIS THEORY EXPLAINED HERE
REGRESSION ANALYSIS THEORY EXPLAINED HEREREGRESSION ANALYSIS THEORY EXPLAINED HERE
REGRESSION ANALYSIS THEORY EXPLAINED HEREShriramKargaonkar
 
Statistics
Statistics Statistics
Statistics KafiPati
 
Statistics-Regression analysis
Statistics-Regression analysisStatistics-Regression analysis
Statistics-Regression analysisRabin BK
 
Regression analysis
Regression analysisRegression analysis
Regression analysisbijuhari
 
Regression.ppt basic introduction of regression with example
Regression.ppt basic introduction of regression with exampleRegression.ppt basic introduction of regression with example
Regression.ppt basic introduction of regression with exampleshivshankarshiva98
 
Regression and Co-Relation
Regression and Co-RelationRegression and Co-Relation
Regression and Co-Relationnuwan udugampala
 
Chapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares RegressionChapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares Regressionnszakir
 
SM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptx
SM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptxSM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptx
SM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptxManjulasingh17
 
CORRELATION-AND-REGRESSION.pdf for human resource
CORRELATION-AND-REGRESSION.pdf for human resourceCORRELATION-AND-REGRESSION.pdf for human resource
CORRELATION-AND-REGRESSION.pdf for human resourceSharon517605
 
TTests.ppt
TTests.pptTTests.ppt
TTests.pptMUzair21
 
Linear regression and correlation analysis ppt @ bec doms
Linear regression and correlation analysis ppt @ bec domsLinear regression and correlation analysis ppt @ bec doms
Linear regression and correlation analysis ppt @ bec domsBabasab Patil
 
correlation.final.ppt (1).pptx
correlation.final.ppt (1).pptxcorrelation.final.ppt (1).pptx
correlation.final.ppt (1).pptxChieWoo1
 

Similar to Cr-and-Regress.ppt (20)

Regression Analysis.pdf
Regression Analysis.pdfRegression Analysis.pdf
Regression Analysis.pdf
 
regression and correlation
regression and correlationregression and correlation
regression and correlation
 
Correlation and Regression
Correlation and Regression Correlation and Regression
Correlation and Regression
 
REGRESSION ANALYSIS THEORY EXPLAINED HERE
REGRESSION ANALYSIS THEORY EXPLAINED HEREREGRESSION ANALYSIS THEORY EXPLAINED HERE
REGRESSION ANALYSIS THEORY EXPLAINED HERE
 
Regression
RegressionRegression
Regression
 
Statistics
Statistics Statistics
Statistics
 
Statistics-Regression analysis
Statistics-Regression analysisStatistics-Regression analysis
Statistics-Regression analysis
 
regression.pptx
regression.pptxregression.pptx
regression.pptx
 
Regression analysis
Regression analysisRegression analysis
Regression analysis
 
Regression.ppt basic introduction of regression with example
Regression.ppt basic introduction of regression with exampleRegression.ppt basic introduction of regression with example
Regression.ppt basic introduction of regression with example
 
Regression and Co-Relation
Regression and Co-RelationRegression and Co-Relation
Regression and Co-Relation
 
Reg
RegReg
Reg
 
Stat
StatStat
Stat
 
Chapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares RegressionChapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares Regression
 
SM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptx
SM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptxSM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptx
SM_d89ccf05-7de1-4a30-a134-3143e9b3bf3f_38.pptx
 
CORRELATION-AND-REGRESSION.pdf for human resource
CORRELATION-AND-REGRESSION.pdf for human resourceCORRELATION-AND-REGRESSION.pdf for human resource
CORRELATION-AND-REGRESSION.pdf for human resource
 
ML Module 3.pdf
ML Module 3.pdfML Module 3.pdf
ML Module 3.pdf
 
TTests.ppt
TTests.pptTTests.ppt
TTests.ppt
 
Linear regression and correlation analysis ppt @ bec doms
Linear regression and correlation analysis ppt @ bec domsLinear regression and correlation analysis ppt @ bec doms
Linear regression and correlation analysis ppt @ bec doms
 
correlation.final.ppt (1).pptx
correlation.final.ppt (1).pptxcorrelation.final.ppt (1).pptx
correlation.final.ppt (1).pptx
 

Recently uploaded

Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxPooja Bhuva
 
FICTIONAL SALESMAN/SALESMAN SNSW 2024.pdf
FICTIONAL SALESMAN/SALESMAN SNSW 2024.pdfFICTIONAL SALESMAN/SALESMAN SNSW 2024.pdf
FICTIONAL SALESMAN/SALESMAN SNSW 2024.pdfPondicherry University
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - Englishneillewis46
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Jisc
 
dusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learningdusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learningMarc Dusseiller Dusjagr
 
Basic Intentional Injuries Health Education
Basic Intentional Injuries Health EducationBasic Intentional Injuries Health Education
Basic Intentional Injuries Health EducationNeilDeclaro1
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxDr. Sarita Anand
 
NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...
NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...
NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...Amil baba
 
Exploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptx
Exploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptxExploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptx
Exploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptxPooja Bhuva
 
AIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.pptAIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.pptNishitharanjan Rout
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxmarlenawright1
 
Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfSherif Taha
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Pooja Bhuva
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...Poonam Aher Patil
 
21st_Century_Skills_Framework_Final_Presentation_2.pptx
21st_Century_Skills_Framework_Final_Presentation_2.pptx21st_Century_Skills_Framework_Final_Presentation_2.pptx
21st_Century_Skills_Framework_Final_Presentation_2.pptxJoelynRubio1
 
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...Nguyen Thanh Tu Collection
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfDr Vijay Vishwakarma
 

Recently uploaded (20)

Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptx
 
FICTIONAL SALESMAN/SALESMAN SNSW 2024.pdf
FICTIONAL SALESMAN/SALESMAN SNSW 2024.pdfFICTIONAL SALESMAN/SALESMAN SNSW 2024.pdf
FICTIONAL SALESMAN/SALESMAN SNSW 2024.pdf
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - English
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
dusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learningdusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learning
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
Basic Intentional Injuries Health Education
Basic Intentional Injuries Health EducationBasic Intentional Injuries Health Education
Basic Intentional Injuries Health Education
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptx
 
NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...
NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...
NO1 Top Black Magic Specialist In Lahore Black magic In Pakistan Kala Ilam Ex...
 
Exploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptx
Exploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptxExploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptx
Exploring_the_Narrative_Style_of_Amitav_Ghoshs_Gun_Island.pptx
 
AIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.pptAIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.ppt
 
Call Girls in Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in  Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7Call Girls in  Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
 
Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdf
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
21st_Century_Skills_Framework_Final_Presentation_2.pptx
21st_Century_Skills_Framework_Final_Presentation_2.pptx21st_Century_Skills_Framework_Final_Presentation_2.pptx
21st_Century_Skills_Framework_Final_Presentation_2.pptx
 
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
 

Cr-and-Regress.ppt

  • 2. Topics Covered:  Is there a relationship between x and y?  What is the strength of this relationship  Pearson’s r  Can we describe this relationship and use this to predict y from x?  Regression  Is the relationship we have described statistically significant?  t test  Relevance to SPM  GLM
  • 3. The relationship between x and y  Correlation: is there a relationship between 2 variables?  Regression: how well a certain independent variable predict dependent variable?  CORRELATION  CAUSATION In order to infer causality: manipulate independent variable and observe effect on dependent variable
  • 4. Scattergrams Y X Y X Y X Y Y Y Positive correlation Negative correlation No correlation
  • 5. Variance vs Covariance  First, a note on your sample:  If you’re wishing to assume that your sample is representative of the general population (RANDOM EFFECTS MODEL), use the degrees of freedom (n – 1) in your calculations of variance or covariance.  But if you’re simply wanting to assess your current sample (FIXED EFFECTS MODEL), substitute n for the degrees of freedom.
  • 6. Variance vs Covariance  Do two variables change together? 1 ) )( ( ) , cov( 1       n y y x x y x i n i i Covariance: • Gives information on the degree to which two variables vary together. • Note how similar the covariance is to variance: the equation simply multiplies x’s error scores by y’s error scores as opposed to squaring x’s error scores. 1 ) ( 2 1 2      n x x S n i i x Variance: • Gives information on variability of a single variable.
  • 7. Covariance  When X and Y : cov (x,y) = pos.  When X and Y : cov (x,y) = neg.  When no constant relationship: cov (x,y) = 0 1 ) )( ( ) , cov( 1       n y y x x y x i n i i
  • 8. Example Covariance x y x xi  y yi  ( x i x  )( y i y  ) 0 3 -3 0 0 2 2 -1 -1 1 3 4 0 1 0 4 0 1 -3 -3 6 6 3 3 9 3  x 3  y  7 75 . 1 4 7 1 )) )( ( ) , cov( 1         n y y x x y x i n i i What does this number tell us? 0 1 2 3 4 5 6 7 0 1 2 3 4 5 6 7
  • 9. Problem with Covariance:  The value obtained by covariance is dependent on the size of the data’s standard deviations: if large, the value will be greater than if small… even if the relationship between x and y is exactly the same in the large versus small standard deviation datasets.
  • 10. Example of how covariance value relies on variance High variance data Low variance data Subject x y x error * y error x y X error * y error 1 101 100 2500 54 53 9 2 81 80 900 53 52 4 3 61 60 100 52 51 1 4 51 50 0 51 50 0 5 41 40 100 50 49 1 6 21 20 900 49 48 4 7 1 0 2500 48 47 9 Mean 51 50 51 50 Sum of x error * y error : 7000 Sum of x error * y error : 28 Covariance: 1166.67 Covariance: 4.67
  • 11. Solution: Pearson’s r  Covariance does not really tell us anything  Solution: standardise this measure  Pearson’s R: standardises the covariance value.  Divides the covariance by the multiplied standard deviations of X and Y: y x xy s s y x r ) , cov( 
  • 12. Pearson’s R continued 1 ) )( ( ) , cov( 1       n y y x x y x i n i i y x i n i i xy s s n y y x x r ) 1 ( ) )( ( 1       1 * 1     n Z Z r n i y x xy i i
  • 13. Limitations of r  When r = 1 or r = -1:  We can predict y from x with certainty  all data points are on a straight line: y = ax + b  r is actually  r = true r of whole population  = estimate of r based on data  r is very sensitive to extreme values: 0 1 2 3 4 5 0 1 2 3 4 5 6 r̂ r̂
  • 14. Regression  Correlation tells you if there is an association between x and y but it doesn’t describe the relationship or allow you to predict one variable from the other.  To do this we need REGRESSION!
  • 15. Best-fit Line = ŷ, predicted value  Aim of linear regression is to fit a straight line, ŷ = ax + b, to data that gives best prediction of y for any value of x  This will be the line that minimises distance between data and fitted line, i.e. the residuals intercept ε ŷ = ax + b ε = residual error = y i , true value slope
  • 16. Least Squares Regression  To find the best line we must minimise the sum of the squares of the residuals (the vertical distances from the data points to our line) Residual (ε) = y - ŷ Sum of squares of residuals = Σ (y – ŷ)2 Model line: ŷ = ax + b  we must find values of a and b that minimise Σ (y – ŷ)2 a = slope, b = intercept
  • 17. Finding b  First we find the value of b that gives the min sum of squares ε ε b b b  Trying different values of b is equivalent to shifting the line up and down the scatter plot
  • 18. Finding a  Now we find the value of a that gives the min sum of squares b b b  Trying out different values of a is equivalent to changing the slope of the line, while b stays constant
  • 19. Minimising sums of squares  Need to minimise Σ(y–ŷ)2  ŷ = ax + b  so need to minimise: Σ(y - ax - b)2  If we plot the sums of squares for all different values of a and b we get a parabola, because it is a squared term  So the min sum of squares is at the bottom of the curve, where the gradient is zero. Values of a and b sums of squares (S) Gradient = 0 min S
  • 20. The maths bit  The min sum of squares is at the bottom of the curve where the gradient = 0  So we can find a and b that give min sum of squares by taking partial derivatives of Σ(y - ax - b)2 with respect to a and b separately  Then we solve these for 0 to give us the values of a and b that give the min sum of squares
  • 21. The solution  Doing this gives the following equations for a and b: a = r sy sx r = correlation coefficient of x and y sy = standard deviation of y sx = standard deviation of x  From you can see that:  A low correlation coefficient gives a flatter slope (small value of a)  Large spread of y, i.e. high standard deviation, results in a steeper slope (high value of a)  Large spread of x, i.e. high standard deviation, results in a flatter slope (high value of a)
  • 22. The solution cont.  Our model equation is ŷ = ax + b  This line must pass through the mean so: y = ax + b b = y – ax  We can put our equation for a into this giving: b = y – ax b = y - r sy sx r = correlation coefficient of x and y sy = standard deviation of y sx = standard deviation of x x  The smaller the correlation, the closer the intercept is to the mean of y
  • 23. Back to the model  If the correlation is zero, we will simply predict the mean of y for every value of x, and our regression line is just a flat straight line crossing the x-axis at y  But this isn’t very useful.  We can calculate the regression line for any data, but the important question is how well does this line fit the data, or how good is it at predicting y from x ŷ = ax + b = r sy sx r sy sx x + y - x r sy sx ŷ = (x – x) + y Rearranges to: a b a a
  • 24. How good is our model?  Total variance of y: sy 2 = ∑(y – y)2 n - 1 SSy dfy =  Variance of predicted y values (ŷ):  Error variance: sŷ 2 = ∑(ŷ – y)2 n - 1 SSpred dfŷ = This is the variance explained by our regression model serror 2 = ∑(y – ŷ)2 n - 2 SSer dfer = This is the variance of the error between our predicted y values and the actual y values, and thus is the variance in y that is NOT explained by the regression model
  • 25.  Total variance = predicted variance + error variance sy 2 = sŷ 2 + ser 2  Conveniently, via some complicated rearranging sŷ 2 = r2 sy 2 r2 = sŷ 2 / sy 2  so r2 is the proportion of the variance in y that is explained by our regression model How good is our model cont.
  • 26. How good is our model cont.  Insert r2 sy 2 into sy 2 = sŷ 2 + ser 2 and rearrange to get: ser 2 = sy 2 – r2sy 2 = sy 2 (1 – r2)  From this we can see that the greater the correlation the smaller the error variance, so the better our prediction
  • 27. Is the model significant?  i.e. do we get a significantly better prediction of y from our regression equation than by just predicting the mean?  F-statistic: F(dfŷ,dfer) = sŷ 2 ser 2 =......= r2 (n - 2)2 1 – r2 complicated rearranging  And it follows that: t(n-2) = r (n - 2) √1 – r2 (because F = t2) So all we need to know are r and n
  • 28. General Linear Model  Linear regression is actually a form of the General Linear Model where the parameters are a, the slope of the line, and b, the intercept. y = ax + b +ε  A General Linear Model is just any model that describes the data in terms of a straight line
  • 29. Multiple regression  Multiple regression is used to determine the effect of a number of independent variables, x1, x2, x3 etc, on a single dependent variable, y  The different x variables are combined in a linear way and each has its own regression coefficient: y = a1x1+ a2x2 +…..+ anxn + b + ε  The a parameters reflect the independent contribution of each independent variable, x, to the value of the dependent variable, y.  i.e. the amount of variance in y that is accounted for by each x variable after all the other x variables have been accounted for
  • 30. SPM  Linear regression is a GLM that models the effect of one independent variable, x, on ONE dependent variable, y  Multiple Regression models the effect of several independent variables, x1, x2 etc, on ONE dependent variable, y  Both are types of General Linear Model  GLM can also allow you to analyse the effects of several independent x variables on several dependent variables, y1, y2, y3 etc, in a linear combination  This is what SPM does and all will be explained next week!