SlideShare a Scribd company logo
1 of 48
Download to read offline
Chapter 9
Random Processes
ENCS6161 - Probability and Stochastic
Processes
Concordia University
Definition of a Random Process
Assume the we have a random experiment with
outcomes w belonging to the sample set S. To each
w ∈ S, we assign a time function X(t, w), t ∈ I, where
I is a time index set: discrete or continuous. X(t, w)
is called a random process.
If w is fixed, X(t, w) is a deterministic time function,
and is called a realization, a sample path, or a
sample function of the random process.
If t = t0 is fixed, X(t0, w) as a function of w, is a
random variable.
A random process is also called a stochastic process.
ENCS6161 – p.1/47
Definition of a Random Process
Example: A random experment has two outcomes
w ∈ {0, 1}. If we assign:
X(t, 0) = A cos t
X(t, 1) = A sin t
where A is a constant. Then X(t, w) is a random
process.
Usually we drop w and write the random process as
X(t).
ENCS6161 – p.2/47
Specifying a Random Process
Joint distribution of time samples
Let X1, · · · , Xn be the samples of X(t, w) obtained at
t1, · · · , tn, i.e. Xi = X(ti, w), then we can use the joint
CDF
FX1···Xn
(x1, · · · , xn) = P[X1 ≤ x1, · · · , Xn ≤ xn]
or the joint pdf fX1···Xn
(x1, · · · , xn) to describe a
random process partially.
Mean function:
mX(t) = E[X(t)] =
Z ∞
−∞
xfX(t)(x)dx
Autocorrelation function
RX(t1, t2) = E[X(t1)X(t2)] =
Z ∞
−∞
Z ∞
−∞
xyfX(t1)X(t2)(x, y)dxdy
ENCS6161 – p.3/47
Specifying a Random Process
Autocovariance function
CX(t1, t2) = E[(X(t1) − mX(t1))(X(t2) − mX(t2))]
= RX(t1, t2) − mX(t1)mX(t2)
a special case:
CX(t, t) = E[(X(t) − mX(t))2
] = V ar[X(t)]
The correlation coefficient
ρX(t1, t2) =
CX(t1, t2)
p
CX(t1, t1)CX(t2, t2)
Mean and autocorrelation functions provide a partial
description of a random process. Only in certain
cases (Gaussian), they can provide a fully
description.
ENCS6161 – p.4/47
Specifying a Random Process
Example: X(t) = A cos(2πt), where A is a random
variable.
mX(t) = E[A cos(2πt)] = E[A] cos(2πt)
RX(t1, t2) = E[A cos(2πt1) · A cos(2πt2)]
= E[A2
] cos(2πt1) cos(2πt2)
CX(t1, t2) = RX(t1, t2) − mX(t1)mX(t2)
= (E[A2
] − E[A]2
) cos(2πt1) cos(2πt2)
= V ar(A) cos(2πt1) cos(2πt2)
ENCS6161 – p.5/47
Specifying a Random Process
Example: X(t) = A cos(wt + Θ), where Θ is uniform in
[0, 2π], A and w are constants.
mX(t) = E[A cos(wt + Θ)]
=
1
2π
Z 2π
0
A cos(wt + θ)dθ = 0
CX(t1, t2) = RX(t1, t2) = A2
E[cos(wt1 + Θ) cos(wt2 + Θ)]
=
A2
2π
Z 2π
0
cos w(t1 − t2) + cos[w(t1 + t2) + θ]
2
dθ
=
A2
2
cos w(t1 − t2)
ENCS6161 – p.6/47
Gaussian Random Processes
A random process X(t) is a Gaussian random
process if for any n, the samples taken at t1, t2, · · · , tn
are jointly Gaussian, i.e. if
X1 = X(t1), · · · , Xn = X(tn)
then
fX1···Xn
(x1, · · · , xn) =
e−1
2
(x−m)T
K−1
(x−m)
(2π)n/2|K|1/2
where m = [mX(t1), · · · , mX(tn)]T and
K =



CX(t1, t1) · · · CX(t1, tn)
· · · · · · · · ·
CX(tn, t1) · · · CX(tn, tn)



ENCS6161 – p.7/47
Multiple Random Processes
To specify joint random processes X(t) and Y (t), we
need to have the pdf of all samples of X(t) and Y (t)
such as X(t1), · · · , X(ti), Y (´
t1), · · · , Y (´
tj) for all i and
j and all choices of t1, · · · , ti, ´
t1, · · · , ´
tj.
The processes X(t) and Y (t) are indepedent if the
random vectors (X(t1), · · · , X(ti)) and
(Y (´
t1), · · · , Y (´
tj)) are independent for all i, j and
t1, · · · , ti, ´
t1, · · · , ´
tj.
ENCS6161 – p.8/47
Multiple Random Processes
The cross-correlation RX,Y (t1, t2) is defined as
RX,Y (t1, t2) = E[X(t1)Y (t2)]
Two processes are orthogonal if
RX,Y (t1, t2) = 0 for all t1 and t2
The cross-covariance
CX,Y (t1, t2) = E[(X(t1) − mX(t1))(Y (t2) − mY (t2))]
= RX,Y (t1, t2) − mX(t1)mY (t2)
X(t) and Y (t) are uncorrelated if
CX,Y (t1, t2) = 0 for all t1 and t2
ENCS6161 – p.9/47
Multiple Random Processes
Example: X(t) = cos(wt + Θ) and Y (t) = sin(wt + Θ),
where Θ is uniform in [0, 2π] and w is a constant.
mX(t) = mY (t) = 0
CX,Y (t1, t2) = RX,Y (t1, t2)
= E[cos(wt1 + Θ) sin(wt2 + Θ)]
= E

−
1
2
sin w(t1 − t2) +
1
2
sin(w(t1 + t2) + 2Θ)

= −
1
2
sin w(t1 − t2)
ENCS6161 – p.10/47
Discrete-Time Random Processes
i.i.d random processes: Xn ∼ fX(xn) then
FX1···Xn
(x1, · · · , xn) = FX(x1) · · · FX(xn)
mX(n) = E[Xn] = m for all n
CX(n1, n2) = E[(Xn1 − m)(Xn2 − m)]
= E[Xn1 − m]E[Xn2 − m] = 0 if n1 6= n2
CX(n, n) = E[(Xn − m)2
] = σ2
⇒ CX(n1, n2) = σ2
δn1,n2
RX(n1, n2) = σ2
δn1,n2 + m2
ENCS6161 – p.11/47
Discrete-Time Random Processes
Example: let Xn be a sequence of i.i.d. Bernoulli
r.v.s. with P(Xi = 1) = p.
mX(n) = p
V ar(Xn) = p(1 − p)
CX(n1, n2) = p(1 − p)δn1,n2
RX(n1, n2) = p(1 − p)δn1,n2 + p2
Example:
Yn = 2Xn − 1, where Xn are i.i.d. Bernoulli r.v.s
Yn =
(
1 with p
−1 with (1 − p)
⇒ mY (n) = 2p − 1, V ar(Yn) = 4p(1 − p)
ENCS6161 – p.12/47
Random Walk
Let Sn = Y1 + · · · + Yn, where Yn are i.i.d. r.v.s. with
P{Yn = 1} = p and P{Yn = −1} = 1 − p. This is a
one-dimensional random walk.
If there are k positive jumps (+1’s) in n trials (n
walks), then there are n − k negative jumps (−1’s).
So Sn = k × 1 + (n − k) × (−1) = 2k − n and
P {Sn = 2k − n} =

n
k

pk
(1 − p)n−k
, k = 0, 1, · · · , n
ENCS6161 – p.13/47
Properties of Random Walk
Independent increment
Let I1 = (n0, n1] and I2 = (n2, n3]. If n1 ≤ n2, I2 and I2
do not overlap. Then the increments on the two
intervals are
Sn1 − Sn0 = Yn0+1 + · · · + Yn1
Sn3 − Sn2 = Yn2+1 + · · · + Yn3
Since they have no Yn’s in common (no overlapping)
and Yn’s are independent.
⇒ Sn1 − Sn0 and Sn3 − Sn2 are independent.
This property is called independent increment.
ENCS6161 – p.14/47
Properties of Random Walk
Stationary increment
Furthermore, if I1 and I2 have the same length, i.e
n1 − n0 = n3 − n2 = m, then the increments Sn1 − Sn0
and Sn3 − Sn2 have the same distribution since they
both are the sum of m i.i.d r.v.s
This means that the increments over interval of the
same length have the same distribution. The process
Sn is said to have stationary increment.
ENCS6161 – p.15/47
Properties of Random Walk
These two properties can be used to find the joint pmf of Sn
at n1, · · · , nk
P[Sn1 = s1, Sn2 = s2, · · · , Snk
= sk]
= P[Sn1 = s1, Sn2 − Sn1 = s2 − s1, · · · , Snk
− Snk−1
= sk − sk−1]
= P[Sn1 = s1]P[Sn2 − Sn1 = s2 − s1] · · · P[Snk
− Snk−1
= sk − sk−1]
(from independent increment)
= P[Sn1 = s1]P[Sn2−n1 = s2 − s1] · · · P[Snk−nk−1
= sk − sk−1]
(from stationary increment)
ENCS6161 – p.16/47
Properties of Random Walk
If Yn are continuous valued r.v.s.
fSn1 ···Snk
(s1, · · · sk)
= fSn1
(s1)fSn2−n1
(s2 − s1) · · · fSnk−nk−1
(sk − sk−1)
e.g., if Yn ∼ N(0, σ2) then
fSn1 ,Sn2
(s1, s2) = fSn1
(s1)fSn2−n1
(s2 − s1)
=
1
√
2πn1σ
e
−
s2
1
2n1σ2
·
1
p
2π(n2 − n1)σ
e
−
(s2−s1)2
2(n2−n1)σ2
ENCS6161 – p.17/47
Sum of i.i.d Processes
If X1, X2, ..., Xn are i.i.d and Sn = X1 + X2 + ... + Xn, we call Sn
the sum process of i.i.d, e.g. random walk is a sum process.
mS(n) = E[Sn] = nE[X] = nm
V ar[Sn] = nV ar[X] = nσ2
Autocovariance
CS(n, k) = E[(Sn − E[Sn])(Sk − E[Sk])]
= E[(Sn − nm)(Sk − km)] = E


n
X
i=1
(Xi − m)
k
X
j=0
(Xj − m)


=
n
X
i=1
k
X
j=0
E[(Xi − m)(Xj − m)] =
n
X
i=1
k
X
j=0
CX(i, j)
=
n
X
i=1
k
X
j=0
σ2
δij = min(n, k)σ2
ENCS6161 – p.18/47
Sum of i.i.d Processes
Example: For random Walk
E[Sn] = nm = n(2p − 1)
V ar[Sn] = nσ2
= 4np(1 − p)
CS(n, k) = min(n, k)4p(1 − p)
ENCS6161 – p.19/47
Continuos Time Random Processes
Poisson Process: a good model for arrival process
N(t) : Number of arrivals in [0, t]
λ : arrival rate (average # of arrivals per time unit)
We divide [0, t] into n subintervals, each with duration
δ = t
n
Assume:
The probability of more than one arrival in a
subinterval is negligible.
Whether or not an event (arrival) occurs in a
subinterval is independent of arrivals in other
subintervals.
So the arrivals in each subinterval are Bernoulli and
they are independent.
ENCS6161 – p.20/47
Poisson Process
Let p = Prob{1 arrival}. Then the average number of
arrivals in [0, t] is
np = λt ⇒ p =
λt
n
The total arrivals in [0, t] ∼ Bionomial(n, p)
P[N(t) = k] =

n
k

pk
(1 − p)k
→
(λt)k
k!
e−λt
when n → ∞.
Stationary increment? Yes
Independent increment? Yes
ENCS6161 – p.21/47
Poisson Process
Inter-arrival Time: Let T be the inter-arrival time.
P{T  t} = P{no arrivals in t seconds}
= P{N(t) = 0} = e−λt
⇒ P{T ≤ t} = 1 − e−λt
fT (t) = λe−λt
, for t ≥ 0
So the inter-arrival time is exponential with mean 1
λ.
ENCS6161 – p.22/47
Random Telegraph Signal
Read on your own
ENCS6161 – p.23/47
Wiener Process
Suppose that the symmetric random walk (p = 1
2)
takes steps of magnitude of h every δ seconds. At
time t , we have n = t
δ jumps.
Xδ(t) = h(D1 + D2 + · · · + Dn) = hSn
where Di are i.i.d random variables taking ±1 with
equal probability.
E[Xδ(t)] = hE[Sn] = 0
V ar[Xδ(t)] = h2
nV ar[Di] = h2
n
ENCS6161 – p.24/47
Wiener Process
If we let h =
√
αδ, where α is a constant and δ → 0
and let the limit of Xδ(t) be X(t), then X(t) is a
continuous-time random process and we have:
E[X(t)] = 0
V ar[X(t)] = lim
δ→0
h2
n = lim
δ→0
(
√
αδ)2 t
δ
= αt
X(t) is called the Wiener process. It is used to model
Brownian motion, the motion of particles suspended
in a fluid that move under the rapid and random
impact of neighbooring particles.
ENCS6161 – p.25/47
Wiener Process
Note that since δ = t
n,
Xδ(t) = hSn =
√
αδSn =
Sn
√
n
√
αt
When δ → 0, n → ∞ and since µD = 0, σD = 1, from
CLT, we have
Sn
√
n
=
Sn − nµD
σD
√
n
∼ N(0, 1)
So the distribution of X(t) follows
X(t) ∼ N(0, αt)
i.e.
fX(t)(x) =
1
√
2παt
e
−x2
2αt
ENCS6161 – p.26/47
Wiener Process
Since Wiener process is a limit of random walk, it
inherits the properties such as independent and
stationary increments. So the joint pdf of X(t) at
t1, t2, · · · , tk (t1  t2  · · ·  tk) will be
fX(t1),X(t2),··· ,X(tk)(x1, x2, · · · , xk)
= fX(t1)(x1)fX(t2−t1)(x2 − x1) · · · fX(tk−tk−1)(xk − xk−1)
=
exp{−1
2[ x2
1
αt1
+ · · · + (xk−xk−1)2
α(tk−tk−1) ]}
p
(2πα)kt1(t2 − t1) · · · (tk − tk−1)
ENCS6161 – p.27/47
Wiener Process
mean funtion: mX(t) = E[X(t)] = 0
auto-covariance: CX(t1, t2) = α min(t1, t2)
Proof:
Xδ(t) = hSn
CXδ
(t1, t2) = h2
CS(n1, n2) (where n1 =
t1
δ
, n2 =
t2
δ
)
= (
√
αδ)2
min(n1, n2)σ2
D
(keep in mind that: σ2
D = 1)
= α min(n1δ, n2δ) = α min(t1, t2)
ENCS6161 – p.28/47
Stationary Random Processes
X(t) is stationary if the joint distribution of any set of
samples does not depend on the placement of the
time origin.
FX(t1),X(t2),··· ,X(tk)(x1, x2, · · · , xk)
= FX(t1+τ),X(t2+τ),··· ,X(tk+τ)(x1, x2, · · · , xk)
for all time shift τ, all k, and all choices of t1, t2, · · · , tk.
X(t) and Y (t) are joint stationary if the joint
distribution of X(t1), X(t2), · · · , X(tk) and
Y (t0
1), Y (t0
2), · · · , Y (t0
j) do not depend on the
placement of the time origin for all k, j and all choices
of t1, t2, · · · , tk and t0
1, t0
2, · · · , t0
j.
ENCS6161 – p.29/47
Stationary Random Processes
First-Order Stationary
FX(t)(x) = FX(t+τ)(x) = FX(x), for all t, τ
⇒ mX(t) = E[X(t)] = m, for all t
V arX(t) = E[(X(t) − m)2
] = σ2
, for all t
Second-Order Stationary
FX(t1)X(t2)(x1, x2) = FX(0)X(t2−t1)(x1, x2), for all t1, t2
⇒ RX(t1, t2) = RX(t1 − t2), for all t1, t2
CX(t1, t2) = CX(t1 − t2), for all t1, t2
The auto-correlation and auto-covariance depend
only on the time difference.
ENCS6161 – p.30/47
Stationary Random Processes
Example:
An i.i.d random process is stationary.
FX(t1),X(t2),··· ,X(tk)(x1, x2, · · · , xk)
= FX(x1)FX(x2) · · · FX(xk)
= FX(t1+τ),X(t2+τ),··· ,X(tk+τ)(x1, x2, · · · , xk)
sum of i.i.d random process Sn = X1 + X2 + · · · + Xn
We know mS(n) = nm and V ar[Sn] = nσ2
⇒ not stationary.
ENCS6161 – p.31/47
Wide-Sense Stationary (WSS)
X(t) is WSS if:
mX(t) = m, for all t
CX(t1, t2) = CX(t1 − t2), for all t1, t2
Let τ = t1 − t2, then CX(t1, t2) = CX(τ).
ENCS6161 – p.32/47
Wide-Sense Stationary (WSS)
Example: Let Xn consists of two interleaved
sequences of independent r.v.s.
For n even: Xn ∈ {+1, −1} with p = 1
2
For n odd: Xn ∈ {1
3, −3} with p = 9
10 and 1
10 resp.
Obviously, Xn is not stationary, since its pmf varies
with n. However,
mX(n) = 0 for all n
CX(i, j) =
(
E[Xi]E[Xj] = 0, i 6= j
E[X2
i ] = 1, i = j
= δi,j
⇒ Xn is WSS.
So stationary ⇒ WSS, WSS ; stationary.
ENCS6161 – p.33/47
Autocorrelation of WSS processes
RX(0) = E[X2(t)], for all t.
RX(0): average power of the process.
RX(τ) is an even function.
RX(τ) = E[X(t + τ)X(t)] = E[X(t)X(t + τ)] = RX(−τ)
RX(τ) is a measure of the rate of change of a r.p.
P{|X(t + τ) − X(t)|  ε} = P{(X(t + τ) − X(t))2
 ε2
}
≤
E[(X(t + τ) − X(t))2]
ε2
(Markov Inequality)
=
2[RX(0) − RX(τ)]
ε2
If RX(τ) is flat ⇒ [RX(0) − RX(τ)] is small ⇒ the
probability of having a large change in X(t) in τ
seconds is small.
ENCS6161 – p.34/47
Autocorrelation of WSS processes
|RX(τ)| ≤ RX(0)
Proof: E[(X(t + τ) ± X(t))2] = 2[RX(0) ± RX(τ)] ≥ 0
⇒ |RX(τ)| ≤ RX(0)
If RX(0) = RX(d), then RX(t) is periodic with period
d, and X(t) is mean square periodic, i.e.,
E[(X(t + d) − X(t))2
] = 0
Proof: read textbook (pg.360). Use the inequality
E[XY ]2 ≤ E[X2]E[Y 2] (from |ρ| ≤ 1, sec.4.7)
If X(t) = m + N(t), where N(t) is a zero-mean
process s.t. RN (τ) → 0, as τ → ∞, then
RX(τ) = E[(m+N(t+τ))(m+N(t))] = m2
+RN (t) → m2
as τ → ∞.
ENCS6161 – p.35/47
Autocorrelation of WSS processes
RX(τ) can have three types of components: (1) a
component that → 0, as τ → ∞, (2) a periodic
component, and (3) a component that due to a non
zero mean.
Example: RX(τ) = e−2α|τ|, RY (τ) = a2
2 cos 2πf0τ
If Z(t) = X(t) + Y (t) + m and assume X, Y are
independent with zero mean, then
RZ(τ) = RX(τ) + RY (τ) + m2
ENCS6161 – p.36/47
WSS Gausian Random Process
If a Gaussian r.p. is WSS, then it is stationary (Strict
Sense Stationary)
Proof:
fX(x) =
exp{−1
2(x − m)T K−1(x − m)}
(2π)
n
2 |K|
1
2
m =



mX(t1)
.
.
.
mX(tn)


 K =



CX(t1, t1) · · · CX(t1, tn)
.
.
.
.
.
.
.
.
.
CX(tn, t1) · · · CX(tn, tn)



If X(t) is WSS, then mX(t1) = mX(t2) = · · · = m,
CX(ti, tj) = CX(ti − tj). So fX(x) does not depend on
the choice of the time origin ⇒ Strict Sense
Stationary.
ENCS6161 – p.37/47
Cyclo Stationary Random Process
Read on your own.
ENCS6161 – p.38/47
Continuity of Random Process
Recall that for X1, X2, · · · , Xn, · · ·
Xn → X in m.s. (mean square) if E[(Xn − X)2] → 0,
as n → ∞
Cauchy Criterion
If E[(Xn − Xm)2] → 0 as n → ∞ and m → ∞, then
{Xn} converges in m.s.
Mean Square Continuity
The r.p. X(t) is continuous at t = t0 in m.s. if
E[(X(t) − X(t0))2
] → 0, as t → t0
We wrtie it as: l.i.m.t→t0 X(t) = X(t0) (limit in the
mean)
ENCS6161 – p.39/47
Continuity of Random Process
Condition for mean square continuity
E[(X(t)−X(t0))2
] = RX(t, t)−RX(t0, t)−RX(t, t0)+RX(t0, t0)
If RX(t1, t2) is continuous (both in t1, t2), at point
(t0, t0), then E[(X(t) − X(t0))2] → 0. So X(t) is
continuous at t0 in m.s. if RX(t1, t2) is continuous at
(t0, t0)
If X(t) is WSS, then:
E[(X(t0 + τ) − X(t0))2
] = 2(RX(0) − RX(τ))
So X(t) is continuous at t0, if RX(τ) is continuous at
τ = 0
ENCS6161 – p.40/47
Continuity of Random Process
If X(t) is continuous at t0 in m.s., then
lim
t→t0
mX(t) = mX(t0)
Proof:
V ar[X(t) − X(t0)] ≥ 0
⇒ E[X(t) − X(t0)]2
≤ E[(X(t) − X(t0))2
] → 0
⇒ (mX(t) − mX(t0))2
→ 0
⇒ mX(t) → mX(t0)
ENCS6161 – p.41/47
Continuity of Random Process
Example: Wiener Process:
RX(t1, t2) = α min(t1, t2)
RX(t1, t2) is continous at (t0, t0) ⇒ X(t) is continuous
at t0 in m.s.
Example: Poisson Process:
CN (t1, t2) = λ min(t1, t2)
RN (t1, t2) = λ min(t1, t2) + λ2
t1t2
N(t) is continuous at t0 in m.s.
Note that for any sample poisson process, there are
infinite number of discontinuities, but N(t) is
continuous at any t0 in m.s.
ENCS6161 – p.42/47
Mean Square Derivative
The mean square derivative X0(t) of the r.p. X(t) is
defined as:
X0
(t) = l.i.m.
ε→0
X(t + ε) − X(t)
ε
provided that
lim
ε→0
E

X(t + ε) − X(t)
ε
− X0
(t)
2
#
= 0
The mean square derivative of X(t) at t exists if
∂2
∂t1∂t2
RX(t1, t2) exists at point (t, t).
Proof: read on your own.
For a Gaussian random process X(t), X0(t) is also
Gaussian
ENCS6161 – p.43/47
Mean Square Derivative
Mean, cross-correlation, and auto-correlation of X0(t)
mX0 (t) =
d
dt
mX(t)
RXX0 (t1, t2) =
∂
∂t2
RX(t1, t2)
R0
X(t1, t2) =
∂2
∂t1∂t2
RX(t1, t2)
When X(t) is WSS,
mX0 (t) = 0
RXX0 (τ) =
∂
∂t2
RX(t1 − t2) = −
d
dτ
RX(τ)
RX0 (τ) =
∂
∂t1

∂
∂t2
RX(t1 − t2)

= −
d2
dτ2
RX(τ)
ENCS6161 – p.44/47
Mean Square Derivative
Example: Wiener Process
RX(t1, t2) = α min(t1, t2) ⇒
∂
∂t2
RX(t1, t2) = αu(t1 − t2)
u(·) is the step function and is discontinuous at
t1 = t2. If we use the delta function,
RX0 (t1, t2) =
∂
∂t1
αu(t1, t2) = αδ(t1 − t2)
Note X0(t) is not physically feasible since
E[X0(t)2] = αδ(0) = ∞, i.e., the signal has infinite
power. When t1 6= t2, RX0 (t1, t2) = 0 ⇒ X0(t1), X0(t2)
uncorrelated (note mX0 (t) = 0 for all t) ⇒ independent
since X0(t) is a Gaussian process.
X0(t) is the so called White Gaussian Noise.
ENCS6161 – p.45/47
Mean Square Integrals
The mean square integral of X(t) form t0 to t:
Y (t) =
R t
t0
X(t0)dt0 exists if the integral
R t
t0
R t
t0
RX(u, v)dudv exists.
The mean and autocorrelation of Y (t)
mY (t) =
Z t
t0
mX(t0
)dt0
RY (t1, t2) =
Z t1
t0
Z t2
t0
RX(u, v)dudv
ENCS6161 – p.46/47
Ergodic Theorems
Time Averages of Random Processes and Ergodic
Theorems.
Read on your own.
ENCS6161 – p.47/47

More Related Content

Similar to ch9.pdf

Scattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysisScattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysisVjekoslavKovac1
 
160511 hasegawa lab_seminar
160511 hasegawa lab_seminar160511 hasegawa lab_seminar
160511 hasegawa lab_seminarTomohiro Koana
 
The Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability DistributionThe Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability DistributionPedro222284
 
Introduction to Diffusion Monte Carlo
Introduction to Diffusion Monte CarloIntroduction to Diffusion Monte Carlo
Introduction to Diffusion Monte CarloClaudio Attaccalite
 
Litvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfLitvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfAlexander Litvinenko
 
lecture about different methods and explainanation.ppt
lecture about different methods and explainanation.pptlecture about different methods and explainanation.ppt
lecture about different methods and explainanation.pptsadafshahbaz7777
 
Scholastic process and explaination lectr14.ppt
Scholastic process and explaination lectr14.pptScholastic process and explaination lectr14.ppt
Scholastic process and explaination lectr14.pptsadafshahbaz7777
 
Implementation of parallel randomized algorithm for skew-symmetric matrix game
Implementation of parallel randomized algorithm for skew-symmetric matrix gameImplementation of parallel randomized algorithm for skew-symmetric matrix game
Implementation of parallel randomized algorithm for skew-symmetric matrix gameAjay Bidyarthy
 
Ph 101-9 QUANTUM MACHANICS
Ph 101-9 QUANTUM MACHANICSPh 101-9 QUANTUM MACHANICS
Ph 101-9 QUANTUM MACHANICSChandan Singh
 
Comparison Theorems for SDEs
Comparison Theorems for SDEs Comparison Theorems for SDEs
Comparison Theorems for SDEs Ilya Gikhman
 
Solution set 3
Solution set 3Solution set 3
Solution set 3慧环 赵
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt msFaeco Bot
 
Fixed Point Theorm In Probabilistic Analysis
Fixed Point Theorm In Probabilistic AnalysisFixed Point Theorm In Probabilistic Analysis
Fixed Point Theorm In Probabilistic Analysisiosrjce
 
A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...OctavianPostavaru
 
Phase-Type Distributions for Finite Interacting Particle Systems
Phase-Type Distributions for Finite Interacting Particle SystemsPhase-Type Distributions for Finite Interacting Particle Systems
Phase-Type Distributions for Finite Interacting Particle SystemsStefan Eng
 
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsRao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsChristian Robert
 
Unit v rpq1
Unit v rpq1Unit v rpq1
Unit v rpq1Babu Rao
 

Similar to ch9.pdf (20)

Scattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysisScattering theory analogues of several classical estimates in Fourier analysis
Scattering theory analogues of several classical estimates in Fourier analysis
 
160511 hasegawa lab_seminar
160511 hasegawa lab_seminar160511 hasegawa lab_seminar
160511 hasegawa lab_seminar
 
The Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability DistributionThe Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability Distribution
 
Introduction to Diffusion Monte Carlo
Introduction to Diffusion Monte CarloIntroduction to Diffusion Monte Carlo
Introduction to Diffusion Monte Carlo
 
Litvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfLitvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdf
 
lecture about different methods and explainanation.ppt
lecture about different methods and explainanation.pptlecture about different methods and explainanation.ppt
lecture about different methods and explainanation.ppt
 
Scholastic process and explaination lectr14.ppt
Scholastic process and explaination lectr14.pptScholastic process and explaination lectr14.ppt
Scholastic process and explaination lectr14.ppt
 
Implementation of parallel randomized algorithm for skew-symmetric matrix game
Implementation of parallel randomized algorithm for skew-symmetric matrix gameImplementation of parallel randomized algorithm for skew-symmetric matrix game
Implementation of parallel randomized algorithm for skew-symmetric matrix game
 
Stochastic Processes - part 1
Stochastic Processes - part 1Stochastic Processes - part 1
Stochastic Processes - part 1
 
Ph 101-9 QUANTUM MACHANICS
Ph 101-9 QUANTUM MACHANICSPh 101-9 QUANTUM MACHANICS
Ph 101-9 QUANTUM MACHANICS
 
Comparison Theorems for SDEs
Comparison Theorems for SDEs Comparison Theorems for SDEs
Comparison Theorems for SDEs
 
Stochastic Processes - part 5
Stochastic Processes - part 5Stochastic Processes - part 5
Stochastic Processes - part 5
 
Solution set 3
Solution set 3Solution set 3
Solution set 3
 
residue
residueresidue
residue
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
Fixed Point Theorm In Probabilistic Analysis
Fixed Point Theorm In Probabilistic AnalysisFixed Point Theorm In Probabilistic Analysis
Fixed Point Theorm In Probabilistic Analysis
 
A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...
 
Phase-Type Distributions for Finite Interacting Particle Systems
Phase-Type Distributions for Finite Interacting Particle SystemsPhase-Type Distributions for Finite Interacting Particle Systems
Phase-Type Distributions for Finite Interacting Particle Systems
 
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsRao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
 
Unit v rpq1
Unit v rpq1Unit v rpq1
Unit v rpq1
 

Recently uploaded

Arduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptArduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptSAURABHKUMAR892774
 
Biology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxBiology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxDeepakSakkari2
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ
 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionDr.Costas Sachpazis
 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfme23b1001
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
 
pipeline in computer architecture design
pipeline in computer architecture  designpipeline in computer architecture  design
pipeline in computer architecture designssuser87fa0c1
 
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube ExchangerStudy on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube ExchangerAnamika Sarkar
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxbritheesh05
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort servicejennyeacort
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 
Effects of rheological properties on mixing
Effects of rheological properties on mixingEffects of rheological properties on mixing
Effects of rheological properties on mixingviprabot1
 
INFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETE
INFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETEINFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETE
INFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETEroselinkalist12
 
EduAI - E learning Platform integrated with AI
EduAI - E learning Platform integrated with AIEduAI - E learning Platform integrated with AI
EduAI - E learning Platform integrated with AIkoyaldeepu123
 
complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...asadnawaz62
 
An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...Chandu841456
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)dollysharma2066
 

Recently uploaded (20)

Arduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptArduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.ppt
 
young call girls in Green Park🔝 9953056974 🔝 escort Service
young call girls in Green Park🔝 9953056974 🔝 escort Serviceyoung call girls in Green Park🔝 9953056974 🔝 escort Service
young call girls in Green Park🔝 9953056974 🔝 escort Service
 
Biology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxBiology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptx
 
POWER SYSTEMS-1 Complete notes examples
POWER SYSTEMS-1 Complete notes  examplesPOWER SYSTEMS-1 Complete notes  examples
POWER SYSTEMS-1 Complete notes examples
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdf
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
 
pipeline in computer architecture design
pipeline in computer architecture  designpipeline in computer architecture  design
pipeline in computer architecture design
 
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube ExchangerStudy on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptx
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
Effects of rheological properties on mixing
Effects of rheological properties on mixingEffects of rheological properties on mixing
Effects of rheological properties on mixing
 
INFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETE
INFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETEINFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETE
INFLUENCE OF NANOSILICA ON THE PROPERTIES OF CONCRETE
 
EduAI - E learning Platform integrated with AI
EduAI - E learning Platform integrated with AIEduAI - E learning Platform integrated with AI
EduAI - E learning Platform integrated with AI
 
complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...
 
An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
 

ch9.pdf

  • 1. Chapter 9 Random Processes ENCS6161 - Probability and Stochastic Processes Concordia University
  • 2. Definition of a Random Process Assume the we have a random experiment with outcomes w belonging to the sample set S. To each w ∈ S, we assign a time function X(t, w), t ∈ I, where I is a time index set: discrete or continuous. X(t, w) is called a random process. If w is fixed, X(t, w) is a deterministic time function, and is called a realization, a sample path, or a sample function of the random process. If t = t0 is fixed, X(t0, w) as a function of w, is a random variable. A random process is also called a stochastic process. ENCS6161 – p.1/47
  • 3. Definition of a Random Process Example: A random experment has two outcomes w ∈ {0, 1}. If we assign: X(t, 0) = A cos t X(t, 1) = A sin t where A is a constant. Then X(t, w) is a random process. Usually we drop w and write the random process as X(t). ENCS6161 – p.2/47
  • 4. Specifying a Random Process Joint distribution of time samples Let X1, · · · , Xn be the samples of X(t, w) obtained at t1, · · · , tn, i.e. Xi = X(ti, w), then we can use the joint CDF FX1···Xn (x1, · · · , xn) = P[X1 ≤ x1, · · · , Xn ≤ xn] or the joint pdf fX1···Xn (x1, · · · , xn) to describe a random process partially. Mean function: mX(t) = E[X(t)] = Z ∞ −∞ xfX(t)(x)dx Autocorrelation function RX(t1, t2) = E[X(t1)X(t2)] = Z ∞ −∞ Z ∞ −∞ xyfX(t1)X(t2)(x, y)dxdy ENCS6161 – p.3/47
  • 5. Specifying a Random Process Autocovariance function CX(t1, t2) = E[(X(t1) − mX(t1))(X(t2) − mX(t2))] = RX(t1, t2) − mX(t1)mX(t2) a special case: CX(t, t) = E[(X(t) − mX(t))2 ] = V ar[X(t)] The correlation coefficient ρX(t1, t2) = CX(t1, t2) p CX(t1, t1)CX(t2, t2) Mean and autocorrelation functions provide a partial description of a random process. Only in certain cases (Gaussian), they can provide a fully description. ENCS6161 – p.4/47
  • 6. Specifying a Random Process Example: X(t) = A cos(2πt), where A is a random variable. mX(t) = E[A cos(2πt)] = E[A] cos(2πt) RX(t1, t2) = E[A cos(2πt1) · A cos(2πt2)] = E[A2 ] cos(2πt1) cos(2πt2) CX(t1, t2) = RX(t1, t2) − mX(t1)mX(t2) = (E[A2 ] − E[A]2 ) cos(2πt1) cos(2πt2) = V ar(A) cos(2πt1) cos(2πt2) ENCS6161 – p.5/47
  • 7. Specifying a Random Process Example: X(t) = A cos(wt + Θ), where Θ is uniform in [0, 2π], A and w are constants. mX(t) = E[A cos(wt + Θ)] = 1 2π Z 2π 0 A cos(wt + θ)dθ = 0 CX(t1, t2) = RX(t1, t2) = A2 E[cos(wt1 + Θ) cos(wt2 + Θ)] = A2 2π Z 2π 0 cos w(t1 − t2) + cos[w(t1 + t2) + θ] 2 dθ = A2 2 cos w(t1 − t2) ENCS6161 – p.6/47
  • 8. Gaussian Random Processes A random process X(t) is a Gaussian random process if for any n, the samples taken at t1, t2, · · · , tn are jointly Gaussian, i.e. if X1 = X(t1), · · · , Xn = X(tn) then fX1···Xn (x1, · · · , xn) = e−1 2 (x−m)T K−1 (x−m) (2π)n/2|K|1/2 where m = [mX(t1), · · · , mX(tn)]T and K =    CX(t1, t1) · · · CX(t1, tn) · · · · · · · · · CX(tn, t1) · · · CX(tn, tn)    ENCS6161 – p.7/47
  • 9. Multiple Random Processes To specify joint random processes X(t) and Y (t), we need to have the pdf of all samples of X(t) and Y (t) such as X(t1), · · · , X(ti), Y (´ t1), · · · , Y (´ tj) for all i and j and all choices of t1, · · · , ti, ´ t1, · · · , ´ tj. The processes X(t) and Y (t) are indepedent if the random vectors (X(t1), · · · , X(ti)) and (Y (´ t1), · · · , Y (´ tj)) are independent for all i, j and t1, · · · , ti, ´ t1, · · · , ´ tj. ENCS6161 – p.8/47
  • 10. Multiple Random Processes The cross-correlation RX,Y (t1, t2) is defined as RX,Y (t1, t2) = E[X(t1)Y (t2)] Two processes are orthogonal if RX,Y (t1, t2) = 0 for all t1 and t2 The cross-covariance CX,Y (t1, t2) = E[(X(t1) − mX(t1))(Y (t2) − mY (t2))] = RX,Y (t1, t2) − mX(t1)mY (t2) X(t) and Y (t) are uncorrelated if CX,Y (t1, t2) = 0 for all t1 and t2 ENCS6161 – p.9/47
  • 11. Multiple Random Processes Example: X(t) = cos(wt + Θ) and Y (t) = sin(wt + Θ), where Θ is uniform in [0, 2π] and w is a constant. mX(t) = mY (t) = 0 CX,Y (t1, t2) = RX,Y (t1, t2) = E[cos(wt1 + Θ) sin(wt2 + Θ)] = E − 1 2 sin w(t1 − t2) + 1 2 sin(w(t1 + t2) + 2Θ) = − 1 2 sin w(t1 − t2) ENCS6161 – p.10/47
  • 12. Discrete-Time Random Processes i.i.d random processes: Xn ∼ fX(xn) then FX1···Xn (x1, · · · , xn) = FX(x1) · · · FX(xn) mX(n) = E[Xn] = m for all n CX(n1, n2) = E[(Xn1 − m)(Xn2 − m)] = E[Xn1 − m]E[Xn2 − m] = 0 if n1 6= n2 CX(n, n) = E[(Xn − m)2 ] = σ2 ⇒ CX(n1, n2) = σ2 δn1,n2 RX(n1, n2) = σ2 δn1,n2 + m2 ENCS6161 – p.11/47
  • 13. Discrete-Time Random Processes Example: let Xn be a sequence of i.i.d. Bernoulli r.v.s. with P(Xi = 1) = p. mX(n) = p V ar(Xn) = p(1 − p) CX(n1, n2) = p(1 − p)δn1,n2 RX(n1, n2) = p(1 − p)δn1,n2 + p2 Example: Yn = 2Xn − 1, where Xn are i.i.d. Bernoulli r.v.s Yn = ( 1 with p −1 with (1 − p) ⇒ mY (n) = 2p − 1, V ar(Yn) = 4p(1 − p) ENCS6161 – p.12/47
  • 14. Random Walk Let Sn = Y1 + · · · + Yn, where Yn are i.i.d. r.v.s. with P{Yn = 1} = p and P{Yn = −1} = 1 − p. This is a one-dimensional random walk. If there are k positive jumps (+1’s) in n trials (n walks), then there are n − k negative jumps (−1’s). So Sn = k × 1 + (n − k) × (−1) = 2k − n and P {Sn = 2k − n} = n k pk (1 − p)n−k , k = 0, 1, · · · , n ENCS6161 – p.13/47
  • 15. Properties of Random Walk Independent increment Let I1 = (n0, n1] and I2 = (n2, n3]. If n1 ≤ n2, I2 and I2 do not overlap. Then the increments on the two intervals are Sn1 − Sn0 = Yn0+1 + · · · + Yn1 Sn3 − Sn2 = Yn2+1 + · · · + Yn3 Since they have no Yn’s in common (no overlapping) and Yn’s are independent. ⇒ Sn1 − Sn0 and Sn3 − Sn2 are independent. This property is called independent increment. ENCS6161 – p.14/47
  • 16. Properties of Random Walk Stationary increment Furthermore, if I1 and I2 have the same length, i.e n1 − n0 = n3 − n2 = m, then the increments Sn1 − Sn0 and Sn3 − Sn2 have the same distribution since they both are the sum of m i.i.d r.v.s This means that the increments over interval of the same length have the same distribution. The process Sn is said to have stationary increment. ENCS6161 – p.15/47
  • 17. Properties of Random Walk These two properties can be used to find the joint pmf of Sn at n1, · · · , nk P[Sn1 = s1, Sn2 = s2, · · · , Snk = sk] = P[Sn1 = s1, Sn2 − Sn1 = s2 − s1, · · · , Snk − Snk−1 = sk − sk−1] = P[Sn1 = s1]P[Sn2 − Sn1 = s2 − s1] · · · P[Snk − Snk−1 = sk − sk−1] (from independent increment) = P[Sn1 = s1]P[Sn2−n1 = s2 − s1] · · · P[Snk−nk−1 = sk − sk−1] (from stationary increment) ENCS6161 – p.16/47
  • 18. Properties of Random Walk If Yn are continuous valued r.v.s. fSn1 ···Snk (s1, · · · sk) = fSn1 (s1)fSn2−n1 (s2 − s1) · · · fSnk−nk−1 (sk − sk−1) e.g., if Yn ∼ N(0, σ2) then fSn1 ,Sn2 (s1, s2) = fSn1 (s1)fSn2−n1 (s2 − s1) = 1 √ 2πn1σ e − s2 1 2n1σ2 · 1 p 2π(n2 − n1)σ e − (s2−s1)2 2(n2−n1)σ2 ENCS6161 – p.17/47
  • 19. Sum of i.i.d Processes If X1, X2, ..., Xn are i.i.d and Sn = X1 + X2 + ... + Xn, we call Sn the sum process of i.i.d, e.g. random walk is a sum process. mS(n) = E[Sn] = nE[X] = nm V ar[Sn] = nV ar[X] = nσ2 Autocovariance CS(n, k) = E[(Sn − E[Sn])(Sk − E[Sk])] = E[(Sn − nm)(Sk − km)] = E   n X i=1 (Xi − m) k X j=0 (Xj − m)   = n X i=1 k X j=0 E[(Xi − m)(Xj − m)] = n X i=1 k X j=0 CX(i, j) = n X i=1 k X j=0 σ2 δij = min(n, k)σ2 ENCS6161 – p.18/47
  • 20. Sum of i.i.d Processes Example: For random Walk E[Sn] = nm = n(2p − 1) V ar[Sn] = nσ2 = 4np(1 − p) CS(n, k) = min(n, k)4p(1 − p) ENCS6161 – p.19/47
  • 21. Continuos Time Random Processes Poisson Process: a good model for arrival process N(t) : Number of arrivals in [0, t] λ : arrival rate (average # of arrivals per time unit) We divide [0, t] into n subintervals, each with duration δ = t n Assume: The probability of more than one arrival in a subinterval is negligible. Whether or not an event (arrival) occurs in a subinterval is independent of arrivals in other subintervals. So the arrivals in each subinterval are Bernoulli and they are independent. ENCS6161 – p.20/47
  • 22. Poisson Process Let p = Prob{1 arrival}. Then the average number of arrivals in [0, t] is np = λt ⇒ p = λt n The total arrivals in [0, t] ∼ Bionomial(n, p) P[N(t) = k] = n k pk (1 − p)k → (λt)k k! e−λt when n → ∞. Stationary increment? Yes Independent increment? Yes ENCS6161 – p.21/47
  • 23. Poisson Process Inter-arrival Time: Let T be the inter-arrival time. P{T t} = P{no arrivals in t seconds} = P{N(t) = 0} = e−λt ⇒ P{T ≤ t} = 1 − e−λt fT (t) = λe−λt , for t ≥ 0 So the inter-arrival time is exponential with mean 1 λ. ENCS6161 – p.22/47
  • 24. Random Telegraph Signal Read on your own ENCS6161 – p.23/47
  • 25. Wiener Process Suppose that the symmetric random walk (p = 1 2) takes steps of magnitude of h every δ seconds. At time t , we have n = t δ jumps. Xδ(t) = h(D1 + D2 + · · · + Dn) = hSn where Di are i.i.d random variables taking ±1 with equal probability. E[Xδ(t)] = hE[Sn] = 0 V ar[Xδ(t)] = h2 nV ar[Di] = h2 n ENCS6161 – p.24/47
  • 26. Wiener Process If we let h = √ αδ, where α is a constant and δ → 0 and let the limit of Xδ(t) be X(t), then X(t) is a continuous-time random process and we have: E[X(t)] = 0 V ar[X(t)] = lim δ→0 h2 n = lim δ→0 ( √ αδ)2 t δ = αt X(t) is called the Wiener process. It is used to model Brownian motion, the motion of particles suspended in a fluid that move under the rapid and random impact of neighbooring particles. ENCS6161 – p.25/47
  • 27. Wiener Process Note that since δ = t n, Xδ(t) = hSn = √ αδSn = Sn √ n √ αt When δ → 0, n → ∞ and since µD = 0, σD = 1, from CLT, we have Sn √ n = Sn − nµD σD √ n ∼ N(0, 1) So the distribution of X(t) follows X(t) ∼ N(0, αt) i.e. fX(t)(x) = 1 √ 2παt e −x2 2αt ENCS6161 – p.26/47
  • 28. Wiener Process Since Wiener process is a limit of random walk, it inherits the properties such as independent and stationary increments. So the joint pdf of X(t) at t1, t2, · · · , tk (t1 t2 · · · tk) will be fX(t1),X(t2),··· ,X(tk)(x1, x2, · · · , xk) = fX(t1)(x1)fX(t2−t1)(x2 − x1) · · · fX(tk−tk−1)(xk − xk−1) = exp{−1 2[ x2 1 αt1 + · · · + (xk−xk−1)2 α(tk−tk−1) ]} p (2πα)kt1(t2 − t1) · · · (tk − tk−1) ENCS6161 – p.27/47
  • 29. Wiener Process mean funtion: mX(t) = E[X(t)] = 0 auto-covariance: CX(t1, t2) = α min(t1, t2) Proof: Xδ(t) = hSn CXδ (t1, t2) = h2 CS(n1, n2) (where n1 = t1 δ , n2 = t2 δ ) = ( √ αδ)2 min(n1, n2)σ2 D (keep in mind that: σ2 D = 1) = α min(n1δ, n2δ) = α min(t1, t2) ENCS6161 – p.28/47
  • 30. Stationary Random Processes X(t) is stationary if the joint distribution of any set of samples does not depend on the placement of the time origin. FX(t1),X(t2),··· ,X(tk)(x1, x2, · · · , xk) = FX(t1+τ),X(t2+τ),··· ,X(tk+τ)(x1, x2, · · · , xk) for all time shift τ, all k, and all choices of t1, t2, · · · , tk. X(t) and Y (t) are joint stationary if the joint distribution of X(t1), X(t2), · · · , X(tk) and Y (t0 1), Y (t0 2), · · · , Y (t0 j) do not depend on the placement of the time origin for all k, j and all choices of t1, t2, · · · , tk and t0 1, t0 2, · · · , t0 j. ENCS6161 – p.29/47
  • 31. Stationary Random Processes First-Order Stationary FX(t)(x) = FX(t+τ)(x) = FX(x), for all t, τ ⇒ mX(t) = E[X(t)] = m, for all t V arX(t) = E[(X(t) − m)2 ] = σ2 , for all t Second-Order Stationary FX(t1)X(t2)(x1, x2) = FX(0)X(t2−t1)(x1, x2), for all t1, t2 ⇒ RX(t1, t2) = RX(t1 − t2), for all t1, t2 CX(t1, t2) = CX(t1 − t2), for all t1, t2 The auto-correlation and auto-covariance depend only on the time difference. ENCS6161 – p.30/47
  • 32. Stationary Random Processes Example: An i.i.d random process is stationary. FX(t1),X(t2),··· ,X(tk)(x1, x2, · · · , xk) = FX(x1)FX(x2) · · · FX(xk) = FX(t1+τ),X(t2+τ),··· ,X(tk+τ)(x1, x2, · · · , xk) sum of i.i.d random process Sn = X1 + X2 + · · · + Xn We know mS(n) = nm and V ar[Sn] = nσ2 ⇒ not stationary. ENCS6161 – p.31/47
  • 33. Wide-Sense Stationary (WSS) X(t) is WSS if: mX(t) = m, for all t CX(t1, t2) = CX(t1 − t2), for all t1, t2 Let τ = t1 − t2, then CX(t1, t2) = CX(τ). ENCS6161 – p.32/47
  • 34. Wide-Sense Stationary (WSS) Example: Let Xn consists of two interleaved sequences of independent r.v.s. For n even: Xn ∈ {+1, −1} with p = 1 2 For n odd: Xn ∈ {1 3, −3} with p = 9 10 and 1 10 resp. Obviously, Xn is not stationary, since its pmf varies with n. However, mX(n) = 0 for all n CX(i, j) = ( E[Xi]E[Xj] = 0, i 6= j E[X2 i ] = 1, i = j = δi,j ⇒ Xn is WSS. So stationary ⇒ WSS, WSS ; stationary. ENCS6161 – p.33/47
  • 35. Autocorrelation of WSS processes RX(0) = E[X2(t)], for all t. RX(0): average power of the process. RX(τ) is an even function. RX(τ) = E[X(t + τ)X(t)] = E[X(t)X(t + τ)] = RX(−τ) RX(τ) is a measure of the rate of change of a r.p. P{|X(t + τ) − X(t)| ε} = P{(X(t + τ) − X(t))2 ε2 } ≤ E[(X(t + τ) − X(t))2] ε2 (Markov Inequality) = 2[RX(0) − RX(τ)] ε2 If RX(τ) is flat ⇒ [RX(0) − RX(τ)] is small ⇒ the probability of having a large change in X(t) in τ seconds is small. ENCS6161 – p.34/47
  • 36. Autocorrelation of WSS processes |RX(τ)| ≤ RX(0) Proof: E[(X(t + τ) ± X(t))2] = 2[RX(0) ± RX(τ)] ≥ 0 ⇒ |RX(τ)| ≤ RX(0) If RX(0) = RX(d), then RX(t) is periodic with period d, and X(t) is mean square periodic, i.e., E[(X(t + d) − X(t))2 ] = 0 Proof: read textbook (pg.360). Use the inequality E[XY ]2 ≤ E[X2]E[Y 2] (from |ρ| ≤ 1, sec.4.7) If X(t) = m + N(t), where N(t) is a zero-mean process s.t. RN (τ) → 0, as τ → ∞, then RX(τ) = E[(m+N(t+τ))(m+N(t))] = m2 +RN (t) → m2 as τ → ∞. ENCS6161 – p.35/47
  • 37. Autocorrelation of WSS processes RX(τ) can have three types of components: (1) a component that → 0, as τ → ∞, (2) a periodic component, and (3) a component that due to a non zero mean. Example: RX(τ) = e−2α|τ|, RY (τ) = a2 2 cos 2πf0τ If Z(t) = X(t) + Y (t) + m and assume X, Y are independent with zero mean, then RZ(τ) = RX(τ) + RY (τ) + m2 ENCS6161 – p.36/47
  • 38. WSS Gausian Random Process If a Gaussian r.p. is WSS, then it is stationary (Strict Sense Stationary) Proof: fX(x) = exp{−1 2(x − m)T K−1(x − m)} (2π) n 2 |K| 1 2 m =    mX(t1) . . . mX(tn)    K =    CX(t1, t1) · · · CX(t1, tn) . . . . . . . . . CX(tn, t1) · · · CX(tn, tn)    If X(t) is WSS, then mX(t1) = mX(t2) = · · · = m, CX(ti, tj) = CX(ti − tj). So fX(x) does not depend on the choice of the time origin ⇒ Strict Sense Stationary. ENCS6161 – p.37/47
  • 39. Cyclo Stationary Random Process Read on your own. ENCS6161 – p.38/47
  • 40. Continuity of Random Process Recall that for X1, X2, · · · , Xn, · · · Xn → X in m.s. (mean square) if E[(Xn − X)2] → 0, as n → ∞ Cauchy Criterion If E[(Xn − Xm)2] → 0 as n → ∞ and m → ∞, then {Xn} converges in m.s. Mean Square Continuity The r.p. X(t) is continuous at t = t0 in m.s. if E[(X(t) − X(t0))2 ] → 0, as t → t0 We wrtie it as: l.i.m.t→t0 X(t) = X(t0) (limit in the mean) ENCS6161 – p.39/47
  • 41. Continuity of Random Process Condition for mean square continuity E[(X(t)−X(t0))2 ] = RX(t, t)−RX(t0, t)−RX(t, t0)+RX(t0, t0) If RX(t1, t2) is continuous (both in t1, t2), at point (t0, t0), then E[(X(t) − X(t0))2] → 0. So X(t) is continuous at t0 in m.s. if RX(t1, t2) is continuous at (t0, t0) If X(t) is WSS, then: E[(X(t0 + τ) − X(t0))2 ] = 2(RX(0) − RX(τ)) So X(t) is continuous at t0, if RX(τ) is continuous at τ = 0 ENCS6161 – p.40/47
  • 42. Continuity of Random Process If X(t) is continuous at t0 in m.s., then lim t→t0 mX(t) = mX(t0) Proof: V ar[X(t) − X(t0)] ≥ 0 ⇒ E[X(t) − X(t0)]2 ≤ E[(X(t) − X(t0))2 ] → 0 ⇒ (mX(t) − mX(t0))2 → 0 ⇒ mX(t) → mX(t0) ENCS6161 – p.41/47
  • 43. Continuity of Random Process Example: Wiener Process: RX(t1, t2) = α min(t1, t2) RX(t1, t2) is continous at (t0, t0) ⇒ X(t) is continuous at t0 in m.s. Example: Poisson Process: CN (t1, t2) = λ min(t1, t2) RN (t1, t2) = λ min(t1, t2) + λ2 t1t2 N(t) is continuous at t0 in m.s. Note that for any sample poisson process, there are infinite number of discontinuities, but N(t) is continuous at any t0 in m.s. ENCS6161 – p.42/47
  • 44. Mean Square Derivative The mean square derivative X0(t) of the r.p. X(t) is defined as: X0 (t) = l.i.m. ε→0 X(t + ε) − X(t) ε provided that lim ε→0 E X(t + ε) − X(t) ε − X0 (t) 2 # = 0 The mean square derivative of X(t) at t exists if ∂2 ∂t1∂t2 RX(t1, t2) exists at point (t, t). Proof: read on your own. For a Gaussian random process X(t), X0(t) is also Gaussian ENCS6161 – p.43/47
  • 45. Mean Square Derivative Mean, cross-correlation, and auto-correlation of X0(t) mX0 (t) = d dt mX(t) RXX0 (t1, t2) = ∂ ∂t2 RX(t1, t2) R0 X(t1, t2) = ∂2 ∂t1∂t2 RX(t1, t2) When X(t) is WSS, mX0 (t) = 0 RXX0 (τ) = ∂ ∂t2 RX(t1 − t2) = − d dτ RX(τ) RX0 (τ) = ∂ ∂t1 ∂ ∂t2 RX(t1 − t2) = − d2 dτ2 RX(τ) ENCS6161 – p.44/47
  • 46. Mean Square Derivative Example: Wiener Process RX(t1, t2) = α min(t1, t2) ⇒ ∂ ∂t2 RX(t1, t2) = αu(t1 − t2) u(·) is the step function and is discontinuous at t1 = t2. If we use the delta function, RX0 (t1, t2) = ∂ ∂t1 αu(t1, t2) = αδ(t1 − t2) Note X0(t) is not physically feasible since E[X0(t)2] = αδ(0) = ∞, i.e., the signal has infinite power. When t1 6= t2, RX0 (t1, t2) = 0 ⇒ X0(t1), X0(t2) uncorrelated (note mX0 (t) = 0 for all t) ⇒ independent since X0(t) is a Gaussian process. X0(t) is the so called White Gaussian Noise. ENCS6161 – p.45/47
  • 47. Mean Square Integrals The mean square integral of X(t) form t0 to t: Y (t) = R t t0 X(t0)dt0 exists if the integral R t t0 R t t0 RX(u, v)dudv exists. The mean and autocorrelation of Y (t) mY (t) = Z t t0 mX(t0 )dt0 RY (t1, t2) = Z t1 t0 Z t2 t0 RX(u, v)dudv ENCS6161 – p.46/47
  • 48. Ergodic Theorems Time Averages of Random Processes and Ergodic Theorems. Read on your own. ENCS6161 – p.47/47