SlideShare a Scribd company logo
1 of 25
Download to read offline
Copyright(© MTS-2002GG): You are free to use and modify these s
Introduction to Time Series AnalysisIntroduction to Time Series Analysis
Gloria González-Rivera
University of California, Riverside
and
Jesús Gonzalo U. Carlos III de Madrid
Spring 2002
• Sample Space: , the set of possible outcomes of some random experiment
• Outcome: , a single element of the Sample Space
• Event: , a subset of the Sample Space
• Field: , the collection of Events we will be considering
• Random Variables: , a function from the Sample Space Ω to a State Space S
• State Space: S, a space containing the possible values of a random variables –common choices are
the integers N, reals R, k-vectors Rk
, complex numbers C, positive reals R+, etc
• Probability: , obeying the three rules that you must very well know
• Distribution: is the Borel sets
(intervals, etc)
Brief Review of ProbabilityBrief Review of Probability
Ω∈ω
}{Ω ω=
Ω⊂E
}E:E Ω⊂{=F
S:Z →Ω
]1,0[:P →F
}RA:A{where,]1,0[: ⊂⊂→µ BB
•Random Vectors: Z= (Z1, Z2 , ..., Zn) is a n-dimensional random vector if its components Z1 , ..., Zn
are one-dimensional real-valued random variables
If we interpret t=1, ..., n as equidistant instants of time, Zt can stand for the outcome of an experiment
at time t . Such a time series may, for example, consists of Toyota share prices Zt at n succeeding days.
The new aspect now, compared to a one-dimensional radnom variable, is that now we can talk about
the dependence structure of the random vector.
•Distribution function FZ of Z : It is the collection of the probabilities
Brief Review (cont)Brief Review (cont)
})nz)(nZ,...,1z)(1Z:({P
)nznZ,...,1z1Z(P)z(ZF
≤ω≤ωω=
≤≤=
We suppose that the exchange rate €/$ at every fixed instant t
between 5p.m and 6p.m. this afternoon is random. Therefore we
can interpret it as a realization Zt(ω) of the random variable Zt, and
so we observe Zt(ω), 5<t<6. In order to make a guess at 6 p.m.
about the exchange rate Z19(ω) at 7 p.m. it is reasonable to look at
the whole evolution of Zt(ω) between 5 and 6 p.m. A
mathematical model describing this evolution is called a stochastic
process.
Stochastic ProcessesStochastic Processes
RZZ tt →Ω:),(ω
Suppose that
(1) For a fixed t
Changing the time index, we can generate several random variables:
)(),.......(),( 21
ωωω nttt ZZZ
(2) For fixed ω
This is just a random variable.
RTZ →:ω This is a realization or sample function
This collection of random variables is called a STOCHASTIC
PROCESSA realization of the stochastic process is called a TIME SERIES
From which a realization is:
nt t tz z z,.... ,2 1
A stochastic process is a collection of time indexed random variables
defined on some space Ω.
),Tt),(tZ()Tt,tZ( Ω∈ω∈ω=∈
Stochastic Processes (cont)Stochastic Processes (cont)
Examples of stochastic processesExamples of stochastic processes
E1: Let the index set be T={1, 2, 3} and let the space of outcomes (Ω) be the possible
outcomes associated with tossing one dice:
Ω={1, 2, 3, ,4 ,5, 6}
Define
Z(t, ω)= t + [value on dice]2
t
Therefore for a particular ω, say ω3={3}, the realization or path would be (10, 20, 30).
Q1: Draw all the different realizations (six) of this stochastic process.
Q2: Think on an economic relevant variable as an stochastic process and write down
an example similar to E1 with it. Specify very clear the sample space and the “rule”
that generates the stochastic process.
E2: A Brownian Motion B=(Bt, t ∈[0, infty]):
• It starts at zero: Bo=0
• It has stationary, independent increments
• For evey t>0, Bt has a normal N(0, t) distribution
• It has continuous sample paths: “no jumps”.
Distribution of a Stochastic ProcessDistribution of a Stochastic Process
In analogy to random variables and random vectors we want to introduce non-random
characteristics of a stochastic process such as its distribution, expectation, etc. and
describe its dependence structure. This is a task much more complicated that the
description of a random vector. Indeed, a non-trivial stochastic process Z=(Zt, t ∈ T)
with infinite index set T is an infinite-dimensional object; it can be inderstood as the
infinite collection of the random variables Zt, t ∈ T. Since the values of Z are functions
on T, the distribution of Z should be defined on subsets of a certain “function space”,
i.e.
P(X ∈ A), A ∈ F,
where F is a collection of suitable subsets of this space of functions. This approach is
possible, but requires advanced mathematics, and so we will try something simpler.
The finite-dimensional distributions (fidis) of the stochastic process Z are the
distributions of the finite-dimensional vectors
(Zt1,..., Ztn), t1, ..., tn ∈T,
for all possible choices of times t1, ..., tn ∈ T and every n ≥ 1.
StationarityStationarity
Consider the joint probability distribution of the collection of
random variables
),...,(),.....,( 221121 nnn ttttttttt zZzZzZPzzzF ≤≤≤=
1st
order stationary process if
ktanyforzFzF ktt ,)()( 111 +=
n-order stationary process if
kttanyforzzFzzF ktkttt ,,),(),( 212121 ++=
kttanyforzzFzzF nktkttt nn
,,).....().....( 111 ++=
Definition.
A process is strongly (strictly) stationary if it is a n-order stationary
process for any n.
2nd
order stationary process if
MomentsMoments
2
t
2
t
)tZ,tZcov(
)2t,1t(
)]ttZ)(ttZ[(E)tZ,tZ(Cov
tdz)tz(f2)ttZ(2)ttZ(E2
t)tZ(Var
tdz)tz(ftZt)tZ(E
21
21
221121
σσ
=ρ
µ−µ−=
µ−=µ−=σ=
=µ=
∫
∫
Moments (cont)Moments (cont)
For strictly stationary
process:
22
σσ
µµ
=
=
t
t
because
µµµ ==→= ++ kttktt zFzF 1111
)()(
provided that ∞< ∞<) ( , ) (2
t tZ E Z E
k
ktkttt
ktkttt
ktttkttt
ttktt
ktkttt
zzzz
zzFzzF
ρρρρ
ρρ
=+=−=
=−=
++=
⇒=
⇒=
++
++
),(),(),(
then,andlet
),(),(
),cov(),cov(
),(),(
21
21
2121
2121
2121
The correlation between any two random variables depends on the
time difference
Weak StationarityWeak Stationarity
A process is said to be n-order weakly stationary if all its joint
moments up to order n exist and are time invariant.
Covariance stationary process (2nd
order weakly stationary):
• constant mean
• constant variance
• covariance function depends on time difference between R.V.
Autocovariance and Autocorrelation FunctionsAutocovariance and Autocorrelation Functions
For a covariance stationary process:
0
2
2
)var()var(
),cov(
),(
)(
)(
γ
γ
σ
γ
ρ
γ
σ
µ
kk
ktt
ktt
k
tsst
t
t
ZZ
ZZ
ZZCov
ZVar
ZE
===
=
=
=
+
+
−
]1,1[k:
(ACF)functionationautocorrel:k
Rk:
functionanceautocovari:k
−→ρ
ρ
→γ
γ
Properties of the autocorrelation functionProperties of the autocorrelation function
1.
2.
3.
1then)var(If 00 == ργ tZ
0
k
1
t,coefficienncorrelatioaisSince
γγρ
ρ
≤⇒≤ kk
ktkt
kktktk
kk
kk
ZZE
ZZE
−−
+−−
−
−
=−−=
=−−=
=
=
γµµ
µµγ
ρρ
γγ
))((
))((since )(
Partial Autocorrelation Function (conditional correlation)Partial Autocorrelation Function (conditional correlation)
This function gives the correlation between two random variables
that are k periods apart when the in-between linear dependence
(between t and t+k ) is removed.
),......|,(bygivenisPACFthe
,variablesrandomtwobeandLet
11 −+++
+
kttktt
ktt
ZZZZ
ZZ
ρ
Motivation Think about a regression model
(without loss of generality, assume that
E(Z)=0)
kjkk......2j2k1j1kj
nsexpectatiotake)2(
jktZktejktZtZkk......jktZ2ktZ2kjktZ1ktZ1kktZjktZ
jktZbymultiply(1)
1jjktZwitheduncorrelatisktewhere
ktetZkk......2ktZ2k1ktZ1kkt
Z
−γφ+−γφ+−γφ=γ
−+++−+φ+−+−+φ+−+−+φ=+−+
−+
≥−++
++φ+−+φ+−+φ=
+
kjkkjkjk −−− ++= ρφρφρφρ ......2211j
Dividing by the variance of the process:
kj ,...2,1=
011
2112
1011
.......
.......
.......
ρφρφρ
ρφρφρ
ρφρφρ
kkkkk
kkkk
kkkk
++=
++=
++=
−
−
−

Yule-Walker
equations
0331322313
1330321312
2331320311
0221212
1220211
1110111
3
2
1
ρφρφρφρ
ρφρφρφρ
ρφρφρφρ
ρφρφρ
ρφρφρ
ρφρφρ
++=
++=
++==
+=
+==
=⇒==
k
k
k
1
1
1
1
1
21
1
22
ρ
ρ
ρρ
ρ
φ =⇒
1
1
1
1
1
12
11
21
312
21
11
33
ρρ
ρρ
ρρ
ρρρ
ρρ
ρρ
φ =⇒
E4: Zt=
Yt if t is even
Yt+1 if t is odd
where Yt is a stationary time series. Is Zt weak stationary?
E5: Define the process
St = X1+ ... + Xn ,
where Xi is iid (0, σ2
). Show that for h>0
Cov (St+h, St) = t σ2
,
and therefore St is not weak stationary.
Examples of stochastic processesExamples of stochastic processes
Examples of stochastic processesExamples of stochastic processes (cont)(cont)
E6: White Noise Process
A sequence of uncorrelated random variables is called a white noise
process. { }
0for0),(
)(
)0(normally)(:
2
≠=
=
==
+ kaaCov
aVar
aEa
ktt
at
aatt
σ
µµ



≠
=
=



≠
=
=



≠
=
=
00
01
00
01
00
0
ationautocorrelandanceAutocovari
2
k
k
k
k
k
k
kk
k
a
k
φ
ρ
σ
γ
. . . .1 2 3 4 k
kρ
Dependence: ErgodicityDependence: Ergodicity
• See Reading 1 from Leo Breiman (1969) “Probability and Stochastic Processes: With
a View Toward Applications”
• We want to allow as much dependence as the Law of Large Numbers (LLN) let us do it
• Stationarity is not enough as the following example shows:
E7: Let {Ut} be a sequence of iid r.v uniformly distributed on [0, 1] and let Z be N(0,1)
independent of {Ut}.
Define Yt=Z+Ut . Then Yt is stationary (why?), but
The problem is that there is too much dependence in the sequence {Yt}. In fact the
correlation between Y1 and Yt is always positive for any value of t.
2
1
ZnY
2
1
)tY(E
no
n
1t
tY
n
1
nY
→−
=→
=
=
∑
Ergodicity for the meanErgodicity for the mean
Need to distinguishing between:
1. Ensemble average 2. Time average
Objective: estimate the mean of the
process
Which estimator is the most appropriate? Ensemble average
Problem: It is impossible to calculate
Under which circumstances we can use the time average?
m
Z
z
m
i
i∑=
= 1
n
Z
z
n
t
t∑=
= 1
{ }tZ
Is the time average an unbiased and consistent estimator of the mean?
)( tZE=µ
Ergodicity for the mean (cont)Ergodicity for the mean (cont)
Reminder. Sufficient conditions for consistency of an estimator.
0)ˆvar(limand)ˆ(lim
T
==
∞→∞→
TT
T
E θθθ
1. Time average is asymptotically unbiased
∑ ∑ ===
t t
t
n
ZE
n
zE µµ
1
)(
1
)(
2. Time average is consistent for the mean
=+++++
+++++++++=
=++=
===
−−−−
−−−
=
−−−
= =
−
= =
∑
∑∑∑∑
)](
)()[(
)(
),cov(
1
)var(
0)2()1(
21011102
0
1
212
0
1 1
2
0
1 1
2
ρρρ
ρρρρρρρ
γ
ρρρ
γ
ρ
γ



nn
nn
n
t
nttt
n
t
n
s
st
n
t
n
s
st
n
n
n
ZZ
n
z
Ergodicity for the mean (cont)Ergodicity for the mean (cont)
∑
∑
∑ ∑
↓↓
→−=
−=−=
∞→∞→
−
−−=
k
k
k
nn
n
nk k
kk
n
k
n
z
n
k
n
kn
n
z
k
0
1
)1(
0
2
0
0
0)1(lim)var(lim
)1()()var(
ρ
ρ
γ
ρ
γ
ρ
γ
A covariance-stationary process is ergodic for the mean if
µ== )(lim tZEzp
A sufficient condition for ergodicity for the mean is
0asisthat
or
0k
k
0
→∞→
∞<∞< ∑∑
∞
=
∞
=
k
k
k
k ρ
ργ
Ergodicity under GaussanityErgodicity under Gaussanity
If { }tZ is a stationary Gaussian process, ∞<∑k
kρ
is sufficient to ensure ergodicity for all moments
Where are We?Where are We?
The Prediction Problem as a Motivating Problem:
Predict Zt+1 given some information set It at time t.
The conditional expectation can be modeled in a parametric way or
in a non-parametric way. We will choose in this course the former.
Parametric models can be linear or non-linear. We will choose in
this course the former way too. Summarizing the models we are
going to study and use in this course will be
Parametric and linear models
]tI|1tZ[E1tZ:Solution
2]1tZ1tZ[EMin
+=+
+−+


Some ProblemsSome Problems
P1: Let {Zt} be a sequence of uncorrelated real-valued variables with zero means and unit variances,
and define the “moving average”
for constants α0, α1, ... , αρ . Show that Y is weak stationary and find its autocovariance function
P2: Show that a Gaussian process is strongly stationary if and only if it is weakly stationary
P3: Let X be a stationary Gaussian process with zero mean, unit variance, and autocovariance function
c. Find the autocovariance functions of the process
itZ
r
0i
itY −
=
α=
∑
}t:3)t(X{3Xand}t:2)t(X{2X ∞<<−∞=∞<<−∞=
Appendix: TransformationsAppendix: Transformations
•Goal: To lead to a more manageable process
•Log transformation reduces certain type of
heteroskedasticity. If we assume µt=E(Xt) and V(Xt) = k µ2
t,
the delta method shows that the variance of the log is roughly
constant:
•Differencing eliminates the trend (not very informative about
the nature of the trend)
•Differencing + Log = Relative Change
k)tZ(Var2)t/1()tZ(log(Var)Z(Var2)('f))Z(f(Var =µ≈⇒µ≈
1tZ
1tZtZ
)
1tZ
1tZtZ
1log()
1tZ
tZ
log()1tZlog()tZlog(
−
−−
≈
−
−−
+=
−
=−−

More Related Content

What's hot

Auto Correlation Presentation
Auto Correlation PresentationAuto Correlation Presentation
Auto Correlation PresentationIrfan Hussain
 
Autocorrelation- Detection- part 1- Durbin-Watson d test
Autocorrelation- Detection- part 1- Durbin-Watson d testAutocorrelation- Detection- part 1- Durbin-Watson d test
Autocorrelation- Detection- part 1- Durbin-Watson d testShilpa Chaudhary
 
Structural VAR: the AB model
Structural VAR: the AB modelStructural VAR: the AB model
Structural VAR: the AB modelAnissa ATMANI
 
Moment Generating Functions
Moment Generating FunctionsMoment Generating Functions
Moment Generating Functionsmathscontent
 
Distributed lag model
Distributed lag modelDistributed lag model
Distributed lag modelPawan Kawan
 
Macro Economics -II Growth model
Macro Economics -II Growth modelMacro Economics -II Growth model
Macro Economics -II Growth modelZegeye Paulos
 
Autocorrelation- Concept, Causes and Consequences
Autocorrelation- Concept, Causes and ConsequencesAutocorrelation- Concept, Causes and Consequences
Autocorrelation- Concept, Causes and ConsequencesShilpa Chaudhary
 
Chapter 07 - Autocorrelation.pptx
Chapter 07 - Autocorrelation.pptxChapter 07 - Autocorrelation.pptx
Chapter 07 - Autocorrelation.pptxFarah Amir
 
Applied econometrics
Applied econometricsApplied econometrics
Applied econometricsWaqas Malik
 
Heteroscedasticity
HeteroscedasticityHeteroscedasticity
HeteroscedasticityMuhammad Ali
 
Lecture notes on Johansen cointegration
Lecture notes on Johansen cointegrationLecture notes on Johansen cointegration
Lecture notes on Johansen cointegrationMoses sichei
 
Heteroscedasticity Remedial Measures.pptx
Heteroscedasticity Remedial Measures.pptxHeteroscedasticity Remedial Measures.pptx
Heteroscedasticity Remedial Measures.pptxPatilDevendra5
 
Econometrics lecture 1st
Econometrics lecture 1stEconometrics lecture 1st
Econometrics lecture 1stIshaq Ahmad
 

What's hot (20)

Auto Correlation Presentation
Auto Correlation PresentationAuto Correlation Presentation
Auto Correlation Presentation
 
Heteroscedasticity
HeteroscedasticityHeteroscedasticity
Heteroscedasticity
 
Econometrics ch12
Econometrics ch12Econometrics ch12
Econometrics ch12
 
Multicollinearity PPT
Multicollinearity PPTMulticollinearity PPT
Multicollinearity PPT
 
Autocorrelation- Detection- part 1- Durbin-Watson d test
Autocorrelation- Detection- part 1- Durbin-Watson d testAutocorrelation- Detection- part 1- Durbin-Watson d test
Autocorrelation- Detection- part 1- Durbin-Watson d test
 
Structural VAR: the AB model
Structural VAR: the AB modelStructural VAR: the AB model
Structural VAR: the AB model
 
Moment Generating Functions
Moment Generating FunctionsMoment Generating Functions
Moment Generating Functions
 
Distributed lag model
Distributed lag modelDistributed lag model
Distributed lag model
 
Macro Economics -II Growth model
Macro Economics -II Growth modelMacro Economics -II Growth model
Macro Economics -II Growth model
 
Autocorrelation- Concept, Causes and Consequences
Autocorrelation- Concept, Causes and ConsequencesAutocorrelation- Concept, Causes and Consequences
Autocorrelation- Concept, Causes and Consequences
 
Chapter 07 - Autocorrelation.pptx
Chapter 07 - Autocorrelation.pptxChapter 07 - Autocorrelation.pptx
Chapter 07 - Autocorrelation.pptx
 
Applied econometrics
Applied econometricsApplied econometrics
Applied econometrics
 
Heteroscedasticity
HeteroscedasticityHeteroscedasticity
Heteroscedasticity
 
Econometrics ch2
Econometrics ch2Econometrics ch2
Econometrics ch2
 
Co-integration
Co-integration Co-integration
Co-integration
 
Lecture notes on Johansen cointegration
Lecture notes on Johansen cointegrationLecture notes on Johansen cointegration
Lecture notes on Johansen cointegration
 
Heteroscedasticity Remedial Measures.pptx
Heteroscedasticity Remedial Measures.pptxHeteroscedasticity Remedial Measures.pptx
Heteroscedasticity Remedial Measures.pptx
 
Autocorrelation (1)
Autocorrelation (1)Autocorrelation (1)
Autocorrelation (1)
 
Econometrics
EconometricsEconometrics
Econometrics
 
Econometrics lecture 1st
Econometrics lecture 1stEconometrics lecture 1st
Econometrics lecture 1st
 

Viewers also liked (8)

Forcasting Techniques
Forcasting TechniquesForcasting Techniques
Forcasting Techniques
 
Matlab tme series benni
Matlab tme series benniMatlab tme series benni
Matlab tme series benni
 
Chap15 time series forecasting & index number
Chap15 time series forecasting & index numberChap15 time series forecasting & index number
Chap15 time series forecasting & index number
 
Princing insurance contracts with R
Princing insurance contracts with RPrincing insurance contracts with R
Princing insurance contracts with R
 
Time Series
Time SeriesTime Series
Time Series
 
Tut2 describing the data
Tut2 describing the dataTut2 describing the data
Tut2 describing the data
 
Tut5 stationarity
Tut5 stationarityTut5 stationarity
Tut5 stationarity
 
time series analysis
time series analysistime series analysis
time series analysis
 

Similar to Introduction - Time Series Analysis

4 stochastic processes
4 stochastic processes4 stochastic processes
4 stochastic processesSolo Hermelin
 
Stability Result of Iterative procedure in normed space
Stability Result of Iterative procedure in normed spaceStability Result of Iterative procedure in normed space
Stability Result of Iterative procedure in normed spaceKomal Goyal
 
digital control Chapter 2 slide
digital control Chapter 2 slidedigital control Chapter 2 slide
digital control Chapter 2 slideasyrafjpk
 
Dcs lec02 - z-transform
Dcs   lec02 - z-transformDcs   lec02 - z-transform
Dcs lec02 - z-transformAmr E. Mohamed
 
Z Transform And Inverse Z Transform - Signal And Systems
Z Transform And Inverse Z Transform - Signal And SystemsZ Transform And Inverse Z Transform - Signal And Systems
Z Transform And Inverse Z Transform - Signal And SystemsMr. RahüL YøGi
 
Polya recurrence
Polya recurrencePolya recurrence
Polya recurrenceBrian Burns
 
The concept of moments of order invariant quantum
The concept of moments of order invariant quantumThe concept of moments of order invariant quantum
The concept of moments of order invariant quantumAlexander Decker
 
Univariate Financial Time Series Analysis
Univariate Financial Time Series AnalysisUnivariate Financial Time Series Analysis
Univariate Financial Time Series AnalysisAnissa ATMANI
 
11.generalized and subset integrated autoregressive moving average bilinear t...
11.generalized and subset integrated autoregressive moving average bilinear t...11.generalized and subset integrated autoregressive moving average bilinear t...
11.generalized and subset integrated autoregressive moving average bilinear t...Alexander Decker
 
Unit-1 DAA_Notes.pdf
Unit-1 DAA_Notes.pdfUnit-1 DAA_Notes.pdf
Unit-1 DAA_Notes.pdfAmayJaiswal4
 
PART VII.3 - Quantum Electrodynamics
PART VII.3 - Quantum ElectrodynamicsPART VII.3 - Quantum Electrodynamics
PART VII.3 - Quantum ElectrodynamicsMaurice R. TREMBLAY
 
Block Cipher vs. Stream Cipher
Block Cipher vs. Stream CipherBlock Cipher vs. Stream Cipher
Block Cipher vs. Stream CipherAmirul Wiramuda
 
Circuit Network Analysis - [Chapter4] Laplace Transform
Circuit Network Analysis - [Chapter4] Laplace TransformCircuit Network Analysis - [Chapter4] Laplace Transform
Circuit Network Analysis - [Chapter4] Laplace TransformSimen Li
 
asymptotic notations i
asymptotic notations iasymptotic notations i
asymptotic notations iAli mahmood
 

Similar to Introduction - Time Series Analysis (20)

4 stochastic processes
4 stochastic processes4 stochastic processes
4 stochastic processes
 
Asymptotic Analysis.ppt
Asymptotic Analysis.pptAsymptotic Analysis.ppt
Asymptotic Analysis.ppt
 
Stability Result of Iterative procedure in normed space
Stability Result of Iterative procedure in normed spaceStability Result of Iterative procedure in normed space
Stability Result of Iterative procedure in normed space
 
digital control Chapter 2 slide
digital control Chapter 2 slidedigital control Chapter 2 slide
digital control Chapter 2 slide
 
Dcs lec02 - z-transform
Dcs   lec02 - z-transformDcs   lec02 - z-transform
Dcs lec02 - z-transform
 
Z Transform And Inverse Z Transform - Signal And Systems
Z Transform And Inverse Z Transform - Signal And SystemsZ Transform And Inverse Z Transform - Signal And Systems
Z Transform And Inverse Z Transform - Signal And Systems
 
Polya recurrence
Polya recurrencePolya recurrence
Polya recurrence
 
The concept of moments of order invariant quantum
The concept of moments of order invariant quantumThe concept of moments of order invariant quantum
The concept of moments of order invariant quantum
 
Univariate Financial Time Series Analysis
Univariate Financial Time Series AnalysisUnivariate Financial Time Series Analysis
Univariate Financial Time Series Analysis
 
11.generalized and subset integrated autoregressive moving average bilinear t...
11.generalized and subset integrated autoregressive moving average bilinear t...11.generalized and subset integrated autoregressive moving average bilinear t...
11.generalized and subset integrated autoregressive moving average bilinear t...
 
Unit-1 DAA_Notes.pdf
Unit-1 DAA_Notes.pdfUnit-1 DAA_Notes.pdf
Unit-1 DAA_Notes.pdf
 
PART VII.3 - Quantum Electrodynamics
PART VII.3 - Quantum ElectrodynamicsPART VII.3 - Quantum Electrodynamics
PART VII.3 - Quantum Electrodynamics
 
Block Cipher vs. Stream Cipher
Block Cipher vs. Stream CipherBlock Cipher vs. Stream Cipher
Block Cipher vs. Stream Cipher
 
Mcqmc talk
Mcqmc talkMcqmc talk
Mcqmc talk
 
residue
residueresidue
residue
 
Lec10
Lec10Lec10
Lec10
 
Lecture5
Lecture5Lecture5
Lecture5
 
Chapter0
Chapter0Chapter0
Chapter0
 
Circuit Network Analysis - [Chapter4] Laplace Transform
Circuit Network Analysis - [Chapter4] Laplace TransformCircuit Network Analysis - [Chapter4] Laplace Transform
Circuit Network Analysis - [Chapter4] Laplace Transform
 
asymptotic notations i
asymptotic notations iasymptotic notations i
asymptotic notations i
 

Recently uploaded

Q3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptx
Q3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptxQ3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptx
Q3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptxArdeniel
 
Physics Serway Jewett 6th edition for Scientists and Engineers
Physics Serway Jewett 6th edition for Scientists and EngineersPhysics Serway Jewett 6th edition for Scientists and Engineers
Physics Serway Jewett 6th edition for Scientists and EngineersAndreaLucarelli
 
Pests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdf
Pests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdfPests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdf
Pests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdfPirithiRaju
 
TORSION IN GASTROPODS- Anatomical event (Zoology)
TORSION IN GASTROPODS- Anatomical event (Zoology)TORSION IN GASTROPODS- Anatomical event (Zoology)
TORSION IN GASTROPODS- Anatomical event (Zoology)chatterjeesoumili50
 
soft skills question paper set for bba ca
soft skills question paper set for bba casoft skills question paper set for bba ca
soft skills question paper set for bba caohsadfeeling
 
Contracts with Interdependent Preferences (2)
Contracts with Interdependent Preferences (2)Contracts with Interdependent Preferences (2)
Contracts with Interdependent Preferences (2)GRAPE
 
Application of Foraminiferal Ecology- Rahul.pptx
Application of Foraminiferal Ecology- Rahul.pptxApplication of Foraminiferal Ecology- Rahul.pptx
Application of Foraminiferal Ecology- Rahul.pptxRahulVishwakarma71547
 
MARSILEA notes in detail for II year Botany.ppt
MARSILEA  notes in detail for II year Botany.pptMARSILEA  notes in detail for II year Botany.ppt
MARSILEA notes in detail for II year Botany.pptaigil2
 
001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...
001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...
001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...marwaahmad357
 
MARKER ASSISTED SELECTION IN CROP IMPROVEMENT
MARKER ASSISTED SELECTION IN CROP IMPROVEMENTMARKER ASSISTED SELECTION IN CROP IMPROVEMENT
MARKER ASSISTED SELECTION IN CROP IMPROVEMENTjipexe1248
 
Pests of tenai_Identification,Binomics_Dr.UPR
Pests of tenai_Identification,Binomics_Dr.UPRPests of tenai_Identification,Binomics_Dr.UPR
Pests of tenai_Identification,Binomics_Dr.UPRPirithiRaju
 
Shiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky Way
Shiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky WayShiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky Way
Shiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky WaySérgio Sacani
 
Role of herbs in hair care Amla and heena.pptx
Role of herbs in hair care  Amla and  heena.pptxRole of herbs in hair care  Amla and  heena.pptx
Role of herbs in hair care Amla and heena.pptxVaishnaviAware
 
Gene transfer in plants agrobacterium.pdf
Gene transfer in plants agrobacterium.pdfGene transfer in plants agrobacterium.pdf
Gene transfer in plants agrobacterium.pdfNetHelix
 
KeyBio pipeline for bioinformatics and data science
KeyBio pipeline for bioinformatics and data scienceKeyBio pipeline for bioinformatics and data science
KeyBio pipeline for bioinformatics and data scienceLayne Sadler
 
Exploration Method’s in Archaeological Studies & Research
Exploration Method’s in Archaeological Studies & ResearchExploration Method’s in Archaeological Studies & Research
Exploration Method’s in Archaeological Studies & ResearchPrachya Adhyayan
 
Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...
Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...
Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...Sérgio Sacani
 
Identification of Superclusters and Their Properties in the Sloan Digital Sky...
Identification of Superclusters and Their Properties in the Sloan Digital Sky...Identification of Superclusters and Their Properties in the Sloan Digital Sky...
Identification of Superclusters and Their Properties in the Sloan Digital Sky...Sérgio Sacani
 

Recently uploaded (20)

Q3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptx
Q3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptxQ3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptx
Q3W4part1-SSSSSSSSSSSSSSSSSSSSSSSSCI.pptx
 
Physics Serway Jewett 6th edition for Scientists and Engineers
Physics Serway Jewett 6th edition for Scientists and EngineersPhysics Serway Jewett 6th edition for Scientists and Engineers
Physics Serway Jewett 6th edition for Scientists and Engineers
 
Pests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdf
Pests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdfPests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdf
Pests of cumbu_Identification, Binomics, Integrated ManagementDr.UPR.pdf
 
TORSION IN GASTROPODS- Anatomical event (Zoology)
TORSION IN GASTROPODS- Anatomical event (Zoology)TORSION IN GASTROPODS- Anatomical event (Zoology)
TORSION IN GASTROPODS- Anatomical event (Zoology)
 
soft skills question paper set for bba ca
soft skills question paper set for bba casoft skills question paper set for bba ca
soft skills question paper set for bba ca
 
Contracts with Interdependent Preferences (2)
Contracts with Interdependent Preferences (2)Contracts with Interdependent Preferences (2)
Contracts with Interdependent Preferences (2)
 
Application of Foraminiferal Ecology- Rahul.pptx
Application of Foraminiferal Ecology- Rahul.pptxApplication of Foraminiferal Ecology- Rahul.pptx
Application of Foraminiferal Ecology- Rahul.pptx
 
MARSILEA notes in detail for II year Botany.ppt
MARSILEA  notes in detail for II year Botany.pptMARSILEA  notes in detail for II year Botany.ppt
MARSILEA notes in detail for II year Botany.ppt
 
001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...
001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...
001 Case Study - Submission Point_c1051231_attempt_2023-11-23-14-08-42_ABS CW...
 
MARKER ASSISTED SELECTION IN CROP IMPROVEMENT
MARKER ASSISTED SELECTION IN CROP IMPROVEMENTMARKER ASSISTED SELECTION IN CROP IMPROVEMENT
MARKER ASSISTED SELECTION IN CROP IMPROVEMENT
 
Pests of tenai_Identification,Binomics_Dr.UPR
Pests of tenai_Identification,Binomics_Dr.UPRPests of tenai_Identification,Binomics_Dr.UPR
Pests of tenai_Identification,Binomics_Dr.UPR
 
Applying Cheminformatics to Develop a Structure Searchable Database of Analyt...
Applying Cheminformatics to Develop a Structure Searchable Database of Analyt...Applying Cheminformatics to Develop a Structure Searchable Database of Analyt...
Applying Cheminformatics to Develop a Structure Searchable Database of Analyt...
 
Shiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky Way
Shiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky WayShiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky Way
Shiva and Shakti: Presumed Proto-Galactic Fragments in the Inner Milky Way
 
Role of herbs in hair care Amla and heena.pptx
Role of herbs in hair care  Amla and  heena.pptxRole of herbs in hair care  Amla and  heena.pptx
Role of herbs in hair care Amla and heena.pptx
 
Data delivery from the US-EPA Center for Computational Toxicology and Exposur...
Data delivery from the US-EPA Center for Computational Toxicology and Exposur...Data delivery from the US-EPA Center for Computational Toxicology and Exposur...
Data delivery from the US-EPA Center for Computational Toxicology and Exposur...
 
Gene transfer in plants agrobacterium.pdf
Gene transfer in plants agrobacterium.pdfGene transfer in plants agrobacterium.pdf
Gene transfer in plants agrobacterium.pdf
 
KeyBio pipeline for bioinformatics and data science
KeyBio pipeline for bioinformatics and data scienceKeyBio pipeline for bioinformatics and data science
KeyBio pipeline for bioinformatics and data science
 
Exploration Method’s in Archaeological Studies & Research
Exploration Method’s in Archaeological Studies & ResearchExploration Method’s in Archaeological Studies & Research
Exploration Method’s in Archaeological Studies & Research
 
Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...
Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...
Legacy Analysis of Dark Matter Annihilation from the Milky Way Dwarf Spheroid...
 
Identification of Superclusters and Their Properties in the Sloan Digital Sky...
Identification of Superclusters and Their Properties in the Sloan Digital Sky...Identification of Superclusters and Their Properties in the Sloan Digital Sky...
Identification of Superclusters and Their Properties in the Sloan Digital Sky...
 

Introduction - Time Series Analysis

  • 1. Copyright(© MTS-2002GG): You are free to use and modify these s Introduction to Time Series AnalysisIntroduction to Time Series Analysis Gloria González-Rivera University of California, Riverside and Jesús Gonzalo U. Carlos III de Madrid Spring 2002
  • 2. • Sample Space: , the set of possible outcomes of some random experiment • Outcome: , a single element of the Sample Space • Event: , a subset of the Sample Space • Field: , the collection of Events we will be considering • Random Variables: , a function from the Sample Space Ω to a State Space S • State Space: S, a space containing the possible values of a random variables –common choices are the integers N, reals R, k-vectors Rk , complex numbers C, positive reals R+, etc • Probability: , obeying the three rules that you must very well know • Distribution: is the Borel sets (intervals, etc) Brief Review of ProbabilityBrief Review of Probability Ω∈ω }{Ω ω= Ω⊂E }E:E Ω⊂{=F S:Z →Ω ]1,0[:P →F }RA:A{where,]1,0[: ⊂⊂→µ BB
  • 3. •Random Vectors: Z= (Z1, Z2 , ..., Zn) is a n-dimensional random vector if its components Z1 , ..., Zn are one-dimensional real-valued random variables If we interpret t=1, ..., n as equidistant instants of time, Zt can stand for the outcome of an experiment at time t . Such a time series may, for example, consists of Toyota share prices Zt at n succeeding days. The new aspect now, compared to a one-dimensional radnom variable, is that now we can talk about the dependence structure of the random vector. •Distribution function FZ of Z : It is the collection of the probabilities Brief Review (cont)Brief Review (cont) })nz)(nZ,...,1z)(1Z:({P )nznZ,...,1z1Z(P)z(ZF ≤ω≤ωω= ≤≤=
  • 4. We suppose that the exchange rate €/$ at every fixed instant t between 5p.m and 6p.m. this afternoon is random. Therefore we can interpret it as a realization Zt(ω) of the random variable Zt, and so we observe Zt(ω), 5<t<6. In order to make a guess at 6 p.m. about the exchange rate Z19(ω) at 7 p.m. it is reasonable to look at the whole evolution of Zt(ω) between 5 and 6 p.m. A mathematical model describing this evolution is called a stochastic process. Stochastic ProcessesStochastic Processes
  • 5. RZZ tt →Ω:),(ω Suppose that (1) For a fixed t Changing the time index, we can generate several random variables: )(),.......(),( 21 ωωω nttt ZZZ (2) For fixed ω This is just a random variable. RTZ →:ω This is a realization or sample function This collection of random variables is called a STOCHASTIC PROCESSA realization of the stochastic process is called a TIME SERIES From which a realization is: nt t tz z z,.... ,2 1 A stochastic process is a collection of time indexed random variables defined on some space Ω. ),Tt),(tZ()Tt,tZ( Ω∈ω∈ω=∈ Stochastic Processes (cont)Stochastic Processes (cont)
  • 6. Examples of stochastic processesExamples of stochastic processes E1: Let the index set be T={1, 2, 3} and let the space of outcomes (Ω) be the possible outcomes associated with tossing one dice: Ω={1, 2, 3, ,4 ,5, 6} Define Z(t, ω)= t + [value on dice]2 t Therefore for a particular ω, say ω3={3}, the realization or path would be (10, 20, 30). Q1: Draw all the different realizations (six) of this stochastic process. Q2: Think on an economic relevant variable as an stochastic process and write down an example similar to E1 with it. Specify very clear the sample space and the “rule” that generates the stochastic process. E2: A Brownian Motion B=(Bt, t ∈[0, infty]): • It starts at zero: Bo=0 • It has stationary, independent increments • For evey t>0, Bt has a normal N(0, t) distribution • It has continuous sample paths: “no jumps”.
  • 7. Distribution of a Stochastic ProcessDistribution of a Stochastic Process In analogy to random variables and random vectors we want to introduce non-random characteristics of a stochastic process such as its distribution, expectation, etc. and describe its dependence structure. This is a task much more complicated that the description of a random vector. Indeed, a non-trivial stochastic process Z=(Zt, t ∈ T) with infinite index set T is an infinite-dimensional object; it can be inderstood as the infinite collection of the random variables Zt, t ∈ T. Since the values of Z are functions on T, the distribution of Z should be defined on subsets of a certain “function space”, i.e. P(X ∈ A), A ∈ F, where F is a collection of suitable subsets of this space of functions. This approach is possible, but requires advanced mathematics, and so we will try something simpler. The finite-dimensional distributions (fidis) of the stochastic process Z are the distributions of the finite-dimensional vectors (Zt1,..., Ztn), t1, ..., tn ∈T, for all possible choices of times t1, ..., tn ∈ T and every n ≥ 1.
  • 8. StationarityStationarity Consider the joint probability distribution of the collection of random variables ),...,(),.....,( 221121 nnn ttttttttt zZzZzZPzzzF ≤≤≤= 1st order stationary process if ktanyforzFzF ktt ,)()( 111 += n-order stationary process if kttanyforzzFzzF ktkttt ,,),(),( 212121 ++= kttanyforzzFzzF nktkttt nn ,,).....().....( 111 ++= Definition. A process is strongly (strictly) stationary if it is a n-order stationary process for any n. 2nd order stationary process if
  • 10. Moments (cont)Moments (cont) For strictly stationary process: 22 σσ µµ = = t t because µµµ ==→= ++ kttktt zFzF 1111 )()( provided that ∞< ∞<) ( , ) (2 t tZ E Z E k ktkttt ktkttt ktttkttt ttktt ktkttt zzzz zzFzzF ρρρρ ρρ =+=−= =−= ++= ⇒= ⇒= ++ ++ ),(),(),( then,andlet ),(),( ),cov(),cov( ),(),( 21 21 2121 2121 2121 The correlation between any two random variables depends on the time difference
  • 11. Weak StationarityWeak Stationarity A process is said to be n-order weakly stationary if all its joint moments up to order n exist and are time invariant. Covariance stationary process (2nd order weakly stationary): • constant mean • constant variance • covariance function depends on time difference between R.V.
  • 12. Autocovariance and Autocorrelation FunctionsAutocovariance and Autocorrelation Functions For a covariance stationary process: 0 2 2 )var()var( ),cov( ),( )( )( γ γ σ γ ρ γ σ µ kk ktt ktt k tsst t t ZZ ZZ ZZCov ZVar ZE === = = = + + − ]1,1[k: (ACF)functionationautocorrel:k Rk: functionanceautocovari:k −→ρ ρ →γ γ
  • 13. Properties of the autocorrelation functionProperties of the autocorrelation function 1. 2. 3. 1then)var(If 00 == ργ tZ 0 k 1 t,coefficienncorrelatioaisSince γγρ ρ ≤⇒≤ kk ktkt kktktk kk kk ZZE ZZE −− +−− − − =−−= =−−= = = γµµ µµγ ρρ γγ ))(( ))((since )(
  • 14. Partial Autocorrelation Function (conditional correlation)Partial Autocorrelation Function (conditional correlation) This function gives the correlation between two random variables that are k periods apart when the in-between linear dependence (between t and t+k ) is removed. ),......|,(bygivenisPACFthe ,variablesrandomtwobeandLet 11 −+++ + kttktt ktt ZZZZ ZZ ρ Motivation Think about a regression model (without loss of generality, assume that E(Z)=0) kjkk......2j2k1j1kj nsexpectatiotake)2( jktZktejktZtZkk......jktZ2ktZ2kjktZ1ktZ1kktZjktZ jktZbymultiply(1) 1jjktZwitheduncorrelatisktewhere ktetZkk......2ktZ2k1ktZ1kkt Z −γφ+−γφ+−γφ=γ −+++−+φ+−+−+φ+−+−+φ=+−+ −+ ≥−++ ++φ+−+φ+−+φ= +
  • 15. kjkkjkjk −−− ++= ρφρφρφρ ......2211j Dividing by the variance of the process: kj ,...2,1= 011 2112 1011 ....... ....... ....... ρφρφρ ρφρφρ ρφρφρ kkkkk kkkk kkkk ++= ++= ++= − − −  Yule-Walker equations 0331322313 1330321312 2331320311 0221212 1220211 1110111 3 2 1 ρφρφρφρ ρφρφρφρ ρφρφρφρ ρφρφρ ρφρφρ ρφρφρ ++= ++= ++== += +== =⇒== k k k 1 1 1 1 1 21 1 22 ρ ρ ρρ ρ φ =⇒ 1 1 1 1 1 12 11 21 312 21 11 33 ρρ ρρ ρρ ρρρ ρρ ρρ φ =⇒
  • 16. E4: Zt= Yt if t is even Yt+1 if t is odd where Yt is a stationary time series. Is Zt weak stationary? E5: Define the process St = X1+ ... + Xn , where Xi is iid (0, σ2 ). Show that for h>0 Cov (St+h, St) = t σ2 , and therefore St is not weak stationary. Examples of stochastic processesExamples of stochastic processes
  • 17. Examples of stochastic processesExamples of stochastic processes (cont)(cont) E6: White Noise Process A sequence of uncorrelated random variables is called a white noise process. { } 0for0),( )( )0(normally)(: 2 ≠= = == + kaaCov aVar aEa ktt at aatt σ µµ    ≠ = =    ≠ = =    ≠ = = 00 01 00 01 00 0 ationautocorrelandanceAutocovari 2 k k k k k k kk k a k φ ρ σ γ . . . .1 2 3 4 k kρ
  • 18. Dependence: ErgodicityDependence: Ergodicity • See Reading 1 from Leo Breiman (1969) “Probability and Stochastic Processes: With a View Toward Applications” • We want to allow as much dependence as the Law of Large Numbers (LLN) let us do it • Stationarity is not enough as the following example shows: E7: Let {Ut} be a sequence of iid r.v uniformly distributed on [0, 1] and let Z be N(0,1) independent of {Ut}. Define Yt=Z+Ut . Then Yt is stationary (why?), but The problem is that there is too much dependence in the sequence {Yt}. In fact the correlation between Y1 and Yt is always positive for any value of t. 2 1 ZnY 2 1 )tY(E no n 1t tY n 1 nY →− =→ = = ∑
  • 19. Ergodicity for the meanErgodicity for the mean Need to distinguishing between: 1. Ensemble average 2. Time average Objective: estimate the mean of the process Which estimator is the most appropriate? Ensemble average Problem: It is impossible to calculate Under which circumstances we can use the time average? m Z z m i i∑= = 1 n Z z n t t∑= = 1 { }tZ Is the time average an unbiased and consistent estimator of the mean? )( tZE=µ
  • 20. Ergodicity for the mean (cont)Ergodicity for the mean (cont) Reminder. Sufficient conditions for consistency of an estimator. 0)ˆvar(limand)ˆ(lim T == ∞→∞→ TT T E θθθ 1. Time average is asymptotically unbiased ∑ ∑ === t t t n ZE n zE µµ 1 )( 1 )( 2. Time average is consistent for the mean =+++++ +++++++++= =++= === −−−− −−− = −−− = = − = = ∑ ∑∑∑∑ )]( )()[( )( ),cov( 1 )var( 0)2()1( 21011102 0 1 212 0 1 1 2 0 1 1 2 ρρρ ρρρρρρρ γ ρρρ γ ρ γ    nn nn n t nttt n t n s st n t n s st n n n ZZ n z
  • 21. Ergodicity for the mean (cont)Ergodicity for the mean (cont) ∑ ∑ ∑ ∑ ↓↓ →−= −=−= ∞→∞→ − −−= k k k nn n nk k kk n k n z n k n kn n z k 0 1 )1( 0 2 0 0 0)1(lim)var(lim )1()()var( ρ ρ γ ρ γ ρ γ A covariance-stationary process is ergodic for the mean if µ== )(lim tZEzp A sufficient condition for ergodicity for the mean is 0asisthat or 0k k 0 →∞→ ∞<∞< ∑∑ ∞ = ∞ = k k k k ρ ργ
  • 22. Ergodicity under GaussanityErgodicity under Gaussanity If { }tZ is a stationary Gaussian process, ∞<∑k kρ is sufficient to ensure ergodicity for all moments
  • 23. Where are We?Where are We? The Prediction Problem as a Motivating Problem: Predict Zt+1 given some information set It at time t. The conditional expectation can be modeled in a parametric way or in a non-parametric way. We will choose in this course the former. Parametric models can be linear or non-linear. We will choose in this course the former way too. Summarizing the models we are going to study and use in this course will be Parametric and linear models ]tI|1tZ[E1tZ:Solution 2]1tZ1tZ[EMin +=+ +−+  
  • 24. Some ProblemsSome Problems P1: Let {Zt} be a sequence of uncorrelated real-valued variables with zero means and unit variances, and define the “moving average” for constants α0, α1, ... , αρ . Show that Y is weak stationary and find its autocovariance function P2: Show that a Gaussian process is strongly stationary if and only if it is weakly stationary P3: Let X be a stationary Gaussian process with zero mean, unit variance, and autocovariance function c. Find the autocovariance functions of the process itZ r 0i itY − = α= ∑ }t:3)t(X{3Xand}t:2)t(X{2X ∞<<−∞=∞<<−∞=
  • 25. Appendix: TransformationsAppendix: Transformations •Goal: To lead to a more manageable process •Log transformation reduces certain type of heteroskedasticity. If we assume µt=E(Xt) and V(Xt) = k µ2 t, the delta method shows that the variance of the log is roughly constant: •Differencing eliminates the trend (not very informative about the nature of the trend) •Differencing + Log = Relative Change k)tZ(Var2)t/1()tZ(log(Var)Z(Var2)('f))Z(f(Var =µ≈⇒µ≈ 1tZ 1tZtZ ) 1tZ 1tZtZ 1log() 1tZ tZ log()1tZlog()tZlog( − −− ≈ − −− += − =−−

Editor's Notes

  1. Add graph to generate intuition
  2. Write graphs showing a process with time-variant mean, and time-variant variance
  3. Make a picture of the autocorrelogram