SlideShare a Scribd company logo
1 of 110
Download to read offline
Advanced Algorithms – COMS31900
Probability recap.
(based on slides by Markus Jalsenius)
Benjamin Sach
Randomness and probability
Probability
The sample space S is the set of outcomes of an experiment.
Probability
The sample space S is the set of outcomes of an experiment.
Roll a die: S = {1, 2, 3, 4, 5, 6}.
EXAMPLES
Probability
The sample space S is the set of outcomes of an experiment.
Roll a die: S = {1, 2, 3, 4, 5, 6}.
EXAMPLES
Flip a coin: S = {H, T}.
Probability
The sample space S is the set of outcomes of an experiment.
Roll a die: S = {1, 2, 3, 4, 5, 6}.
EXAMPLES
Flip a coin: S = {H, T}.
Amount of money you can win when playing some lottery:
S = {£0, £10, £100, £1000, £10, 000, £100, 000}.
Probability
The sample space S is the set of outcomes of an experiment.
Roll a die: S = {1, 2, 3, 4, 5, 6}.
EXAMPLES
Flip a coin: S = {H, T}.
Amount of money you can win when playing some lottery:
S = {£0, £10, £100, £1000, £10, 000, £100, 000}.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Probability
The sample space S is the set of outcomes of an experiment.
Roll a die: S = {1, 2, 3, 4, 5, 6}.
EXAMPLES
Flip a coin: S = {H, T}.
Amount of money you can win when playing some lottery:
S = {£0, £10, £100, £1000, £10, 000, £100, 000}.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Probability
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Probability
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Roll a die: S = {1, 2, 3, 4, 5, 6}.
EXAMPLE
Pr(1) = Pr(2) = Pr(3) = Pr(4) = Pr(5) = Pr(6) = 1
6 .
Probability
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Probability
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Flip a coin: S = {H, T}.
Pr(H) = Pr(T) = 1
2 .
EXAMPLE
Probability
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Probability
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
Amount of money you can win when playing some lottery:
Pr(£0) = 0.9, Pr(£10) = 0.08, . . . , Pr(£100, 000) = 0.0001.
EXAMPLE
S = {£0, £10, £100, £1000, £10, 000, £100, 000}.
Probability
The sample space is not necessarily finite.
Probability
The sample space is not necessarily finite.
EXAMPLE
Flip a coin until first tail shows up
Probability
The sample space is not necessarily finite.
Flip a coin until first tail shows up:
EXAMPLE
S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
Probability
The sample space is not necessarily finite.
Flip a coin until first tail shows up:
Pr(“It takes n coin flips”) = 1
2
n
, and
EXAMPLE
S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
Probability
The sample space is not necessarily finite.
Flip a coin until first tail shows up:
Pr(“It takes n coin flips”) = 1
2
n
, and
EXAMPLE
S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
∞
n=1
1
2
n
Probability
The sample space is not necessarily finite.
Flip a coin until first tail shows up:
Pr(“It takes n coin flips”) = 1
2
n
, and
EXAMPLE
S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
∞
n=1
1
2
n
= 1
2 + 1
4 + 1
8 + 1
16 . . .
Probability
The sample space is not necessarily finite.
Flip a coin until first tail shows up:
Pr(“It takes n coin flips”) = 1
2
n
, and
EXAMPLE
S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
∞
n=1
1
2
n
= 1
2 + 1
4 + 1
8 + 1
16 . . . = 1
Event
An event is a subset V of the sample space S.
Event
An event is a subset V of the sample space S.
The probability of event V happening, denoted Pr(V ), is
Pr(V ) =
x∈V
Pr(x).
Event
An event is a subset V of the sample space S.
EXAMPLE
The probability of event V happening, denoted Pr(V ), is
Pr(V ) =
x∈V
Pr(x).
Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH}
For each x ∈ S, Pr(x) = 1
8
Event
An event is a subset V of the sample space S.
EXAMPLE
The probability of event V happening, denoted Pr(V ), is
Pr(V ) =
x∈V
Pr(x).
Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH}
For each x ∈ S, Pr(x) = 1
8
Define V to be the event “the first and last coin flips are the same”
Event
An event is a subset V of the sample space S.
EXAMPLE
The probability of event V happening, denoted Pr(V ), is
Pr(V ) =
x∈V
Pr(x).
Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH}
For each x ∈ S, Pr(x) = 1
8
Define V to be the event “the first and last coin flips are the same”
in other words, V = {HHH, HTH, THT, TTT}
Event
An event is a subset V of the sample space S.
EXAMPLE
The probability of event V happening, denoted Pr(V ), is
Pr(V ) =
x∈V
Pr(x).
Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH}
For each x ∈ S, Pr(x) = 1
8
Define V to be the event “the first and last coin flips are the same”
in other words, V = {HHH, HTH, THT, TTT}
What is Pr(V )?
Event
An event is a subset V of the sample space S.
Pr(V ) = Pr(HHH) + Pr(HTH) + Pr(THT) + Pr(TTT) = 4 × 1
8 = 1
2 .
EXAMPLE
The probability of event V happening, denoted Pr(V ), is
Pr(V ) =
x∈V
Pr(x).
Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH}
For each x ∈ S, Pr(x) = 1
8
Define V to be the event “the first and last coin flips are the same”
in other words, V = {HHH, HTH, THT, TTT}
What is Pr(V )?
Random variable
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
sum over all values of x such that Y (x) = y
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
sum over all values of x such that Y (x) = y
What is Pr(Y = 2)?
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
sum over all values of x such that Y (x) = y
What is Pr(Y = 2)?
Pr(Y = 2) =
x∈{HH,TT}
Pr(x) =
1
4
+
1
4
=
1
2
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
sum over all values of x such that Y (x) = y
What is Pr(Y = 2)?
Pr(Y = 2) =
x∈{HH,TT}
Pr(x) =
1
4
+
1
4
=
1
2
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
sum over all values of x such that Y (x) = y
What is Pr(Y = 2)?
Pr(Y = 2) =
x∈{HH,TT}
Pr(x) =
1
4
+
1
4
=
1
2
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
Pr(Y = 2) = 1
2
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
sum over all values of x such that Y (x) = y
What is Pr(Y = 2)?
Pr(Y = 2) =
x∈{HH,TT}
Pr(x) =
1
4
+
1
4
=
1
2
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
Pr(Y = 2) = 1
2
EXAMPLE
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
Pr(Y = 2) = 1
2
The expected value (the mean) of a r.v. Y ,
EXAMPLE
denoted E(Y ), is
E(Y ) =
x∈S
Y (x)·Pr(x).
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
Random variable
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
Two coin flips.
H H
H T
T H
T T
2
1
5
2
S Y
Pr(Y = 2) = 1
2
The expected value (the mean) of a r.v. Y ,
E(Y ) = 2 · 1
2 + 1 · 1
4 + 5 · 1
4 = 7
2
EXAMPLE
denoted E(Y ), is
E(Y ) =
x∈S
Y (x)·Pr(x).
A random variable (r.v.) Y over sample space S is a function S → R
i.e. it maps each outcome x ∈ S to some real number Y (x).
Pr(Y = y) = Pr(x).
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
THEOREM (Linearity of expectation)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
(regardless of whether the random variables are independent or not.)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
random variable
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 1: (without the theorem)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 1: (without the theorem)
The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
E(Y ) = x∈S Y (x) · Pr(x) = 1
36 x∈S Y (x) =
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 1: (without the theorem)
The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
E(Y ) = x∈S Y (x) · Pr(x) = 1
36 x∈S Y (x) =
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 1: (without the theorem)
The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
E(Y ) = x∈S Y (x) · Pr(x) = 1
36 x∈S Y (x) =
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 1: (without the theorem)
The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
1
36 (1 · 2 + 2 · 3 + 3 · 4 + · · · + 1 · 12) = 7
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
E(Y ) = x∈S Y (x) · Pr(x) = 1
36 x∈S Y (x) =
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 1: (without the theorem)
The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
1
36 (1 · 2 + 2 · 3 + 3 · 4 + · · · + 1 · 12) = 7
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 2: (with the theorem)
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 2: (with the theorem)
Let the r.v. Y1 be the value of the first die and Y2 the value of the second
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 2: (with the theorem)
Let the r.v. Y1 be the value of the first die and Y2 the value of the second
E(Y1) = E(Y2) = 3.5
Linearity of expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Linearity of expectation always holds,
THEOREM (Linearity of expectation)
EXAMPLE
(regardless of whether the random variables are independent or not.)
Roll two dice. Let the r.v. Y be the sum of the values.
What is E(Y )?
Approach 2: (with the theorem)
Let the r.v. Y1 be the value of the first die and Y2 the value of the second
E(Y1) = E(Y2) = 3.5
so E(Y ) = E(Y1 + Y2) = E(Y1) + E(Y2) = 7
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
(usually referred to by the letter I)
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
(usually referred to by the letter I)
Fact: E(I) = 0 · Pr(I = 0) + 1 · Pr(I = 1) = Pr(I = 1).
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Often an indicator r.v. I is associated with an event such that
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Pr(Ij = 1) = 21
36 = 7
12 . (by counting the outcomes)
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Pr(Ij = 1) = 21
36 = 7
12 . (by counting the outcomes)
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
E
n
j=2
Ij =
n
j=2
E(Ij) =
n
j=2
Pr(Ij = 1) = (n − 1) ·
7
12
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Pr(Ij = 1) = 21
36 = 7
12 . (by counting the outcomes)
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
E
n
j=2
Ij =
n
j=2
E(Ij) =
n
j=2
Pr(Ij = 1) = (n − 1) ·
7
12
Linearity of Expectation
Let Y1, Y2, . . . , Yk be k random variables. Then
E
k
i=1
Yi =
k
i=1
E(Yi)
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Pr(Ij = 1) = 21
36 = 7
12 . (by counting the outcomes)
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
E
n
j=2
Ij =
n
j=2
E(Ij) =
n
j=2
Pr(Ij = 1) = (n − 1) ·
7
12
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Pr(Ij = 1) = 21
36 = 7
12 . (by counting the outcomes)
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
E
n
j=2
Ij =
n
j=2
E(Ij) =
n
j=2
Pr(Ij = 1) = (n − 1) ·
7
12
Indicator random variables
An indicator random variable is a r.v. that can only be 0 or 1.
Roll a die n times.
Pr(Ij = 1) = 21
36 = 7
12 . (by counting the outcomes)
Often an indicator r.v. I is associated with an event such that
Indicator random variables and linearity of expectation work great together!
EXAMPLE
(usually referred to by the letter I)
Fact: E(I) = Pr(I = 1).
I = 1 if the event happens (and I = 0 otherwise).
How many rolls do we expect to show a value
that is at least the value of the previous roll?
For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll
is at least the value of the previous roll (and Ij = 0 otherwise)
E
n
j=2
Ij =
n
j=2
E(Ij) =
n
j=2
Pr(Ij = 1) = (n − 1) ·
7
12
Markov’s inequality
EXAMPLE
Suppose that the average (mean) speed on the motorway is 60 mph.
Markov’s inequality
It then follows that at most
EXAMPLE
Suppose that the average (mean) speed on the motorway is 60 mph.
Markov’s inequality
It then follows that at most
EXAMPLE
Suppose that the average (mean) speed on the motorway is 60 mph.
1
2 of all cars drive at least 120 mph,
Markov’s inequality
It then follows that at most
EXAMPLE
. . . otherwise the mean must be higher than 60 mph. (a contradiction)
Suppose that the average (mean) speed on the motorway is 60 mph.
1
2 of all cars drive at least 120 mph,
Markov’s inequality
It then follows that at most
EXAMPLE
. . . otherwise the mean must be higher than 60 mph. (a contradiction)
Suppose that the average (mean) speed on the motorway is 60 mph.
2
3 of all cars drive at least 90 mph,
Markov’s inequality
It then follows that at most
If X is a non-negative r.v., then for all a > 0,
Pr(X ≥ a) ≤
E(X)
a
.
THEOREM (Markov’s inequality)
EXAMPLE
. . . otherwise the mean must be higher than 60 mph. (a contradiction)
Suppose that the average (mean) speed on the motorway is 60 mph.
2
3 of all cars drive at least 90 mph,
Markov’s inequality
It then follows that at most
If X is a non-negative r.v., then for all a > 0,
Pr(X ≥ a) ≤
E(X)
a
.
From the example above:
Pr(speed of a random car ≥ 120 mph) ≤ 60
120 = 1
2 ,
Pr(speed of a random car ≥ 90mph) ≤ 60
90 = 2
3 .
EXAMPLE
THEOREM (Markov’s inequality)
EXAMPLE
. . . otherwise the mean must be higher than 60 mph. (a contradiction)
Suppose that the average (mean) speed on the motorway is 60 mph.
2
3 of all cars drive at least 90 mph,
Markov’s inequality
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
Markov’s inequality
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
otherwise Ij = 0.
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . . Fact: E(I) = Pr(I = 1).
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
By Markov’s inequality (recall: Pr(X ≥ a) ≤
E(X)
a ),
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
By Markov’s inequality (recall: Pr(X ≥ a) ≤
E(X)
a ),
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Pr(5 or more people leaving with their own hats) ≤ 1
5 ,
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
By Markov’s inequality (recall: Pr(X ≥ a) ≤
E(X)
a ),
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Pr(5 or more people leaving with their own hats) ≤ 1
5 ,
Pr(at least 1 person leaving with their own hat) ≤ 1
1 = 1.
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
By Markov’s inequality (recall: Pr(X ≥ a) ≤
E(X)
a ),
(sometimes Markov’s inequality is not particularly informative)
EXAMPLE
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Pr(5 or more people leaving with their own hats) ≤ 1
5 ,
Pr(at least 1 person leaving with their own hat) ≤ 1
1 = 1.
Markov’s inequality
For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat,
By Markov’s inequality (recall: Pr(X ≥ a) ≤
E(X)
a ),
(sometimes Markov’s inequality is not particularly informative)
EXAMPLE
In fact, here it can be shown that as n → ∞, the probability that at least
one person leaves with their own hat is 1 − 1
e .
n people go to a party, leaving their hats at the door.
Each person leaves with a random hat.
How many people leave with their own hat?
E
n
j=1
Ij =
n
j=1
E(Ij) =
n
j=1
Pr(Ij = 1) = n·
1
n
= 1.
otherwise Ij = 0.
By linearity of expectation. . .
Pr(5 or more people leaving with their own hats) ≤ 1
5 ,
Pr(at least 1 person leaving with their own hat) ≤ 1
1 = 1.
Markov’s inequality
If X is a non-negative r.v. that only takes integer values, then
Pr(X > 0) = Pr(X ≥ 1) ≤ E(X) .
COROLLARY
For an indicator r.v. I, the bound is tight (=), as Pr(I > 0) = E(I).
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
THEOREM (union bound)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
THEOREM (union bound)
This is the probability at least one of the events happens
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
THEOREM (union bound)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Let the r.v. X = k
j=1 Ij be the number of events that happen.
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Pr k
j=1 Vj = Pr(X >0) ≤ E(X) = E( k
j=1 Ij) = k
j=1 E(Ij)
Let the r.v. X = k
j=1 Ij be the number of events that happen.
= k
j=1 Pr(Vj)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
by previous
Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Pr k
j=1 Vj = Pr(X >0) ≤ E(X) = E( k
j=1 Ij) = k
j=1 E(Ij)
Let the r.v. X = k
j=1 Ij be the number of events that happen.
= k
j=1 Pr(Vj)
Markov corollary
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
by previous
Linearity of expectation
Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Pr k
j=1 Vj = Pr(X >0) ≤ E(X) = E( k
j=1 Ij) = k
j=1 E(Ij)
Let the r.v. X = k
j=1 Ij be the number of events that happen.
= k
j=1 Pr(Vj)
Markov corollary
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
by previous
Linearity of expectation
Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0.
THEOREM (union bound)
PROOF
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Pr k
j=1 Vj = Pr(X >0) ≤ E(X) = E( k
j=1 Ij) = k
j=1 E(Ij)
Let the r.v. X = k
j=1 Ij be the number of events that happen.
= k
j=1 Pr(Vj)
Markov corollary
E(I) = Pr(I = 1)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
EXAMPLE
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
EXAMPLE
We define two events: V1 = {3, 4}
V2 = {1, 2, 3}
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1
3 + 1
2 = 5
6
EXAMPLE
We define two events: V1 = {3, 4}
V2 = {1, 2, 3}
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
1
2
S
V1
V2
4
6
3
5
Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1
3 + 1
2 = 5
6
EXAMPLE
We define two events: V1 = {3, 4}
V2 = {1, 2, 3}
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
1
2
S
V1
V2
4
6
3
5
Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1
3 + 1
2 = 5
6
EXAMPLE
We define two events: V1 = {3, 4}
V2 = {1, 2, 3}
in fact, Pr(V1 ∪ V2) = 2
3
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
1
2
S
V1
V2
4
6
3
5
Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1
3 + 1
2 = 5
6
EXAMPLE
We define two events: V1 = {3, 4}
V2 = {1, 2, 3}
in fact, Pr(V1 ∪ V2) = 2
3 (3 was ‘double counted’)
Union bound
Let V1, . . . , Vk be k events. Then
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
This bound is tight (=) when the events are all disjoint.
THEOREM (union bound)
(Vi and Vj are disjoint iff Vi ∩ Vj is empty)
S = {1, . . . , 6} is the set of outcomes of a die roll.
1
2
S
V1
V2
4
6
3
5
Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1
3 + 1
2 = 5
6
EXAMPLE
We define two events: V1 = {3, 4}
V2 = {1, 2, 3}
in fact, Pr(V1 ∪ V2) = 2
3 (3 was ‘double counted’)
Typically the union bound is used when each Pr(Vi) is much smaller than k.
Summary
The sample space S is the set of outcomes of an experiment.
For x ∈ S, the probability of x, written Pr(x),
such that x∈S Pr(x) = 1.
is a real number between 0 and 1,
An event is a subset V of the sample space S, Pr(V ) = x∈V Pr(x)
The probability of Y taking value y is
{x ∈ S st. Y(x) = y}
A random variable (r.v.) Y is a function which maps x ∈ S to S(x) ∈ R
Pr(Y = y) = Pr(x).
The expected value (the mean) of Y is E(Y ) =
x∈S
Y (x)·Pr(x).
An indicator random variable is a r.v. that can only be 0 or 1.
Fact: E(I) = Pr(I = 1).
Let V1, . . . , Vk be k events then,
THEOREM (union bound)
Pr
k
i=1
Vi ≤
k
i=1
Pr(Vi).
If X is a non-negative r.v., then for all a > 0,
THEOREM (Markov’s inequality)
Pr(X ≥ a) ≤
E(X)
a
.
Let Y1,Y2,...,Yk be k random variables then,
E
k
i=1
Yi =
k
i=1
E(Yi)
THEOREM (Linearity of expectation)

More Related Content

What's hot

Lecture 2 predicates quantifiers and rules of inference
Lecture 2 predicates quantifiers and rules of inferenceLecture 2 predicates quantifiers and rules of inference
Lecture 2 predicates quantifiers and rules of inferenceasimnawaz54
 
Predicates and Quantifiers
Predicates and QuantifiersPredicates and Quantifiers
Predicates and Quantifiersblaircomp2003
 
Lesson 16: Derivatives of Exponential and Logarithmic Functions
Lesson 16: Derivatives of Exponential and Logarithmic FunctionsLesson 16: Derivatives of Exponential and Logarithmic Functions
Lesson 16: Derivatives of Exponential and Logarithmic FunctionsMatthew Leingang
 
Deep generative model.pdf
Deep generative model.pdfDeep generative model.pdf
Deep generative model.pdfHyungjoo Cho
 
Integration by Parts, Part 1
Integration by Parts, Part 1Integration by Parts, Part 1
Integration by Parts, Part 1Pablo Antuna
 
Advanced matlab codigos matematicos
Advanced matlab codigos matematicosAdvanced matlab codigos matematicos
Advanced matlab codigos matematicosKmilo Bolaños
 
Chapter 4: Decision theory and Bayesian analysis
Chapter 4: Decision theory and Bayesian analysisChapter 4: Decision theory and Bayesian analysis
Chapter 4: Decision theory and Bayesian analysisChristian Robert
 
Interpolation techniques - Background and implementation
Interpolation techniques - Background and implementationInterpolation techniques - Background and implementation
Interpolation techniques - Background and implementationQuasar Chunawala
 
28 mac laurin expansions x
28 mac laurin expansions x28 mac laurin expansions x
28 mac laurin expansions xmath266
 
Contribution of Fixed Point Theorem in Quasi Metric Spaces
Contribution of Fixed Point Theorem in Quasi Metric SpacesContribution of Fixed Point Theorem in Quasi Metric Spaces
Contribution of Fixed Point Theorem in Quasi Metric SpacesAM Publications,India
 
Largedictionaries handout
Largedictionaries handoutLargedictionaries handout
Largedictionaries handoutcsedays
 
“Going bananas with recursion schemes for fixed point data types”
“Going bananas with recursion schemes for fixed point data types”“Going bananas with recursion schemes for fixed point data types”
“Going bananas with recursion schemes for fixed point data types”Pawel Szulc
 

What's hot (17)

Lecture 2 predicates quantifiers and rules of inference
Lecture 2 predicates quantifiers and rules of inferenceLecture 2 predicates quantifiers and rules of inference
Lecture 2 predicates quantifiers and rules of inference
 
Predicates and Quantifiers
Predicates and QuantifiersPredicates and Quantifiers
Predicates and Quantifiers
 
ScalaMeter 2014
ScalaMeter 2014ScalaMeter 2014
ScalaMeter 2014
 
Lesson 16: Derivatives of Exponential and Logarithmic Functions
Lesson 16: Derivatives of Exponential and Logarithmic FunctionsLesson 16: Derivatives of Exponential and Logarithmic Functions
Lesson 16: Derivatives of Exponential and Logarithmic Functions
 
Deep generative model.pdf
Deep generative model.pdfDeep generative model.pdf
Deep generative model.pdf
 
Scala Parallel Collections
Scala Parallel CollectionsScala Parallel Collections
Scala Parallel Collections
 
Integration by Parts, Part 1
Integration by Parts, Part 1Integration by Parts, Part 1
Integration by Parts, Part 1
 
QMC: Operator Splitting Workshop, Stochastic Block-Coordinate Fixed Point Alg...
QMC: Operator Splitting Workshop, Stochastic Block-Coordinate Fixed Point Alg...QMC: Operator Splitting Workshop, Stochastic Block-Coordinate Fixed Point Alg...
QMC: Operator Splitting Workshop, Stochastic Block-Coordinate Fixed Point Alg...
 
Advanced matlab codigos matematicos
Advanced matlab codigos matematicosAdvanced matlab codigos matematicos
Advanced matlab codigos matematicos
 
Course notes2summer2012
Course notes2summer2012Course notes2summer2012
Course notes2summer2012
 
Chapter 4: Decision theory and Bayesian analysis
Chapter 4: Decision theory and Bayesian analysisChapter 4: Decision theory and Bayesian analysis
Chapter 4: Decision theory and Bayesian analysis
 
Interpolation techniques - Background and implementation
Interpolation techniques - Background and implementationInterpolation techniques - Background and implementation
Interpolation techniques - Background and implementation
 
28 mac laurin expansions x
28 mac laurin expansions x28 mac laurin expansions x
28 mac laurin expansions x
 
Contribution of Fixed Point Theorem in Quasi Metric Spaces
Contribution of Fixed Point Theorem in Quasi Metric SpacesContribution of Fixed Point Theorem in Quasi Metric Spaces
Contribution of Fixed Point Theorem in Quasi Metric Spaces
 
Largedictionaries handout
Largedictionaries handoutLargedictionaries handout
Largedictionaries handout
 
Numerical Methods 3
Numerical Methods 3Numerical Methods 3
Numerical Methods 3
 
“Going bananas with recursion schemes for fixed point data types”
“Going bananas with recursion schemes for fixed point data types”“Going bananas with recursion schemes for fixed point data types”
“Going bananas with recursion schemes for fixed point data types”
 

Viewers also liked

Approximation Algorithms Part One: Constant factor approximations
Approximation Algorithms Part One: Constant factor approximationsApproximation Algorithms Part One: Constant factor approximations
Approximation Algorithms Part One: Constant factor approximationsBenjamin Sach
 
Approximation Algorithms Part Four: APTAS
Approximation Algorithms Part Four: APTASApproximation Algorithms Part Four: APTAS
Approximation Algorithms Part Four: APTASBenjamin Sach
 
Line Segment Intersections
Line Segment IntersectionsLine Segment Intersections
Line Segment IntersectionsBenjamin Sach
 
Basic Concept Of Probability
Basic Concept Of ProbabilityBasic Concept Of Probability
Basic Concept Of Probabilityguest45a926
 
Probability Powerpoint
Probability PowerpointProbability Powerpoint
Probability Powerpointspike2904
 
PROBABILITY
PROBABILITYPROBABILITY
PROBABILITYVIV13
 

Viewers also liked (6)

Approximation Algorithms Part One: Constant factor approximations
Approximation Algorithms Part One: Constant factor approximationsApproximation Algorithms Part One: Constant factor approximations
Approximation Algorithms Part One: Constant factor approximations
 
Approximation Algorithms Part Four: APTAS
Approximation Algorithms Part Four: APTASApproximation Algorithms Part Four: APTAS
Approximation Algorithms Part Four: APTAS
 
Line Segment Intersections
Line Segment IntersectionsLine Segment Intersections
Line Segment Intersections
 
Basic Concept Of Probability
Basic Concept Of ProbabilityBasic Concept Of Probability
Basic Concept Of Probability
 
Probability Powerpoint
Probability PowerpointProbability Powerpoint
Probability Powerpoint
 
PROBABILITY
PROBABILITYPROBABILITY
PROBABILITY
 

Similar to Probability Recap

Random variables
Random variablesRandom variables
Random variablesMenglinLiu1
 
CPSC531-Probability.pptx
CPSC531-Probability.pptxCPSC531-Probability.pptx
CPSC531-Probability.pptxRidaIrfan10
 
Estimation Theory, PhD Course, Ghent University, Belgium
Estimation Theory, PhD Course, Ghent University, BelgiumEstimation Theory, PhD Course, Ghent University, Belgium
Estimation Theory, PhD Course, Ghent University, BelgiumStijn De Vuyst
 
Qt random variables notes
Qt random variables notesQt random variables notes
Qt random variables notesRohan Bhatkar
 
Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)
Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)
Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)jemille6
 
Communication Theory - Random Process.pdf
Communication Theory - Random Process.pdfCommunication Theory - Random Process.pdf
Communication Theory - Random Process.pdfRajaSekaran923497
 
Probability distribution for Dummies
Probability distribution for DummiesProbability distribution for Dummies
Probability distribution for DummiesBalaji P
 
Probability and statistics - Discrete Random Variables and Probability Distri...
Probability and statistics - Discrete Random Variables and Probability Distri...Probability and statistics - Discrete Random Variables and Probability Distri...
Probability and statistics - Discrete Random Variables and Probability Distri...Asma CHERIF
 
Discrete probability
Discrete probabilityDiscrete probability
Discrete probabilityRanjan Kumar
 
Random Variables and Distributions
Random Variables and DistributionsRandom Variables and Distributions
Random Variables and Distributionshussein zayed
 
random variables-descriptive and contincuous
random variables-descriptive and contincuousrandom variables-descriptive and contincuous
random variables-descriptive and contincuousar9530
 
Chapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability DistributionsChapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability DistributionsJasonTagapanGulla
 

Similar to Probability Recap (20)

Random Variable
Random Variable Random Variable
Random Variable
 
U unit7 ssb
U unit7 ssbU unit7 ssb
U unit7 ssb
 
Random variables
Random variablesRandom variables
Random variables
 
CPSC531-Probability.pptx
CPSC531-Probability.pptxCPSC531-Probability.pptx
CPSC531-Probability.pptx
 
Course material mca
Course material   mcaCourse material   mca
Course material mca
 
5 random variables
5 random variables5 random variables
5 random variables
 
Estimation Theory, PhD Course, Ghent University, Belgium
Estimation Theory, PhD Course, Ghent University, BelgiumEstimation Theory, PhD Course, Ghent University, Belgium
Estimation Theory, PhD Course, Ghent University, Belgium
 
Probability[1]
Probability[1]Probability[1]
Probability[1]
 
Qt random variables notes
Qt random variables notesQt random variables notes
Qt random variables notes
 
Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)
Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)
Mayo Slides: Part I Meeting #2 (Phil 6334/Econ 6614)
 
Communication Theory - Random Process.pdf
Communication Theory - Random Process.pdfCommunication Theory - Random Process.pdf
Communication Theory - Random Process.pdf
 
Probability distribution for Dummies
Probability distribution for DummiesProbability distribution for Dummies
Probability distribution for Dummies
 
Probability and statistics - Discrete Random Variables and Probability Distri...
Probability and statistics - Discrete Random Variables and Probability Distri...Probability and statistics - Discrete Random Variables and Probability Distri...
Probability and statistics - Discrete Random Variables and Probability Distri...
 
Probability Theory 9
Probability Theory 9Probability Theory 9
Probability Theory 9
 
Ch5
Ch5Ch5
Ch5
 
Probability
ProbabilityProbability
Probability
 
Discrete probability
Discrete probabilityDiscrete probability
Discrete probability
 
Random Variables and Distributions
Random Variables and DistributionsRandom Variables and Distributions
Random Variables and Distributions
 
random variables-descriptive and contincuous
random variables-descriptive and contincuousrandom variables-descriptive and contincuous
random variables-descriptive and contincuous
 
Chapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability DistributionsChapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability Distributions
 

More from Benjamin Sach

Approximation Algorithms Part Three: (F)PTAS
Approximation Algorithms Part Three: (F)PTASApproximation Algorithms Part Three: (F)PTAS
Approximation Algorithms Part Three: (F)PTASBenjamin Sach
 
Approximation Algorithms Part Two: More Constant factor approximations
Approximation Algorithms Part Two: More Constant factor approximationsApproximation Algorithms Part Two: More Constant factor approximations
Approximation Algorithms Part Two: More Constant factor approximationsBenjamin Sach
 
Orthogonal Range Searching
Orthogonal Range SearchingOrthogonal Range Searching
Orthogonal Range SearchingBenjamin Sach
 
Pattern Matching Part Two: k-mismatches
Pattern Matching Part Two: k-mismatchesPattern Matching Part Two: k-mismatches
Pattern Matching Part Two: k-mismatchesBenjamin Sach
 
Pattern Matching Part Three: Hamming Distance
Pattern Matching Part Three: Hamming DistancePattern Matching Part Three: Hamming Distance
Pattern Matching Part Three: Hamming DistanceBenjamin Sach
 
Lowest Common Ancestor
Lowest Common AncestorLowest Common Ancestor
Lowest Common AncestorBenjamin Sach
 
Range Minimum Queries
Range Minimum QueriesRange Minimum Queries
Range Minimum QueriesBenjamin Sach
 
Pattern Matching Part Two: Suffix Arrays
Pattern Matching Part Two: Suffix ArraysPattern Matching Part Two: Suffix Arrays
Pattern Matching Part Two: Suffix ArraysBenjamin Sach
 
Pattern Matching Part One: Suffix Trees
Pattern Matching Part One: Suffix TreesPattern Matching Part One: Suffix Trees
Pattern Matching Part One: Suffix TreesBenjamin Sach
 
Minimum Spanning Trees (via Disjoint Sets)
Minimum Spanning Trees (via Disjoint Sets)Minimum Spanning Trees (via Disjoint Sets)
Minimum Spanning Trees (via Disjoint Sets)Benjamin Sach
 
Shortest Paths Part 1: Priority Queues and Dijkstra's Algorithm
Shortest Paths Part 1: Priority Queues and Dijkstra's AlgorithmShortest Paths Part 1: Priority Queues and Dijkstra's Algorithm
Shortest Paths Part 1: Priority Queues and Dijkstra's AlgorithmBenjamin Sach
 
Depth First Search and Breadth First Search
Depth First Search and Breadth First SearchDepth First Search and Breadth First Search
Depth First Search and Breadth First SearchBenjamin Sach
 
Shortest Paths Part 2: Negative Weights and All-pairs
Shortest Paths Part 2: Negative Weights and All-pairsShortest Paths Part 2: Negative Weights and All-pairs
Shortest Paths Part 2: Negative Weights and All-pairsBenjamin Sach
 
Self-balancing Trees and Skip Lists
Self-balancing Trees and Skip ListsSelf-balancing Trees and Skip Lists
Self-balancing Trees and Skip ListsBenjamin Sach
 

More from Benjamin Sach (16)

Approximation Algorithms Part Three: (F)PTAS
Approximation Algorithms Part Three: (F)PTASApproximation Algorithms Part Three: (F)PTAS
Approximation Algorithms Part Three: (F)PTAS
 
Approximation Algorithms Part Two: More Constant factor approximations
Approximation Algorithms Part Two: More Constant factor approximationsApproximation Algorithms Part Two: More Constant factor approximations
Approximation Algorithms Part Two: More Constant factor approximations
 
Orthogonal Range Searching
Orthogonal Range SearchingOrthogonal Range Searching
Orthogonal Range Searching
 
Pattern Matching Part Two: k-mismatches
Pattern Matching Part Two: k-mismatchesPattern Matching Part Two: k-mismatches
Pattern Matching Part Two: k-mismatches
 
Pattern Matching Part Three: Hamming Distance
Pattern Matching Part Three: Hamming DistancePattern Matching Part Three: Hamming Distance
Pattern Matching Part Three: Hamming Distance
 
Lowest Common Ancestor
Lowest Common AncestorLowest Common Ancestor
Lowest Common Ancestor
 
Range Minimum Queries
Range Minimum QueriesRange Minimum Queries
Range Minimum Queries
 
Pattern Matching Part Two: Suffix Arrays
Pattern Matching Part Two: Suffix ArraysPattern Matching Part Two: Suffix Arrays
Pattern Matching Part Two: Suffix Arrays
 
Pattern Matching Part One: Suffix Trees
Pattern Matching Part One: Suffix TreesPattern Matching Part One: Suffix Trees
Pattern Matching Part One: Suffix Trees
 
Bloom Filters
Bloom FiltersBloom Filters
Bloom Filters
 
Dynamic Programming
Dynamic ProgrammingDynamic Programming
Dynamic Programming
 
Minimum Spanning Trees (via Disjoint Sets)
Minimum Spanning Trees (via Disjoint Sets)Minimum Spanning Trees (via Disjoint Sets)
Minimum Spanning Trees (via Disjoint Sets)
 
Shortest Paths Part 1: Priority Queues and Dijkstra's Algorithm
Shortest Paths Part 1: Priority Queues and Dijkstra's AlgorithmShortest Paths Part 1: Priority Queues and Dijkstra's Algorithm
Shortest Paths Part 1: Priority Queues and Dijkstra's Algorithm
 
Depth First Search and Breadth First Search
Depth First Search and Breadth First SearchDepth First Search and Breadth First Search
Depth First Search and Breadth First Search
 
Shortest Paths Part 2: Negative Weights and All-pairs
Shortest Paths Part 2: Negative Weights and All-pairsShortest Paths Part 2: Negative Weights and All-pairs
Shortest Paths Part 2: Negative Weights and All-pairs
 
Self-balancing Trees and Skip Lists
Self-balancing Trees and Skip ListsSelf-balancing Trees and Skip Lists
Self-balancing Trees and Skip Lists
 

Recently uploaded

भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,Virag Sontakke
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaVirag Sontakke
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Celine George
 
Hybridoma Technology ( Production , Purification , and Application )
Hybridoma Technology  ( Production , Purification , and Application  ) Hybridoma Technology  ( Production , Purification , and Application  )
Hybridoma Technology ( Production , Purification , and Application ) Sakshi Ghasle
 
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxEPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxRaymartEstabillo3
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17Celine George
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Blooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docxBlooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docxUnboundStockton
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptxENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptxAnaBeatriceAblay2
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdfssuser54595a
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 

Recently uploaded (20)

भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of India
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
 
Hybridoma Technology ( Production , Purification , and Application )
Hybridoma Technology  ( Production , Purification , and Application  ) Hybridoma Technology  ( Production , Purification , and Application  )
Hybridoma Technology ( Production , Purification , and Application )
 
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxEPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Blooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docxBlooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docx
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 
Staff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSDStaff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSD
 
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptxENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 

Probability Recap

  • 1. Advanced Algorithms – COMS31900 Probability recap. (based on slides by Markus Jalsenius) Benjamin Sach
  • 3. Probability The sample space S is the set of outcomes of an experiment.
  • 4. Probability The sample space S is the set of outcomes of an experiment. Roll a die: S = {1, 2, 3, 4, 5, 6}. EXAMPLES
  • 5. Probability The sample space S is the set of outcomes of an experiment. Roll a die: S = {1, 2, 3, 4, 5, 6}. EXAMPLES Flip a coin: S = {H, T}.
  • 6. Probability The sample space S is the set of outcomes of an experiment. Roll a die: S = {1, 2, 3, 4, 5, 6}. EXAMPLES Flip a coin: S = {H, T}. Amount of money you can win when playing some lottery: S = {£0, £10, £100, £1000, £10, 000, £100, 000}.
  • 7. Probability The sample space S is the set of outcomes of an experiment. Roll a die: S = {1, 2, 3, 4, 5, 6}. EXAMPLES Flip a coin: S = {H, T}. Amount of money you can win when playing some lottery: S = {£0, £10, £100, £1000, £10, 000, £100, 000}. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1,
  • 8. Probability The sample space S is the set of outcomes of an experiment. Roll a die: S = {1, 2, 3, 4, 5, 6}. EXAMPLES Flip a coin: S = {H, T}. Amount of money you can win when playing some lottery: S = {£0, £10, £100, £1000, £10, 000, £100, 000}. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
  • 9. Probability The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
  • 10. Probability The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1] Roll a die: S = {1, 2, 3, 4, 5, 6}. EXAMPLE Pr(1) = Pr(2) = Pr(3) = Pr(4) = Pr(5) = Pr(6) = 1 6 .
  • 11. Probability The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
  • 12. Probability The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1] Flip a coin: S = {H, T}. Pr(H) = Pr(T) = 1 2 . EXAMPLE
  • 13. Probability The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1]
  • 14. Probability The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, Pr is ‘just’ a function which maps each x ∈ S to Pr(x) ∈ [0, 1] Amount of money you can win when playing some lottery: Pr(£0) = 0.9, Pr(£10) = 0.08, . . . , Pr(£100, 000) = 0.0001. EXAMPLE S = {£0, £10, £100, £1000, £10, 000, £100, 000}.
  • 15. Probability The sample space is not necessarily finite.
  • 16. Probability The sample space is not necessarily finite. EXAMPLE Flip a coin until first tail shows up
  • 17. Probability The sample space is not necessarily finite. Flip a coin until first tail shows up: EXAMPLE S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
  • 18. Probability The sample space is not necessarily finite. Flip a coin until first tail shows up: Pr(“It takes n coin flips”) = 1 2 n , and EXAMPLE S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }.
  • 19. Probability The sample space is not necessarily finite. Flip a coin until first tail shows up: Pr(“It takes n coin flips”) = 1 2 n , and EXAMPLE S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }. ∞ n=1 1 2 n
  • 20. Probability The sample space is not necessarily finite. Flip a coin until first tail shows up: Pr(“It takes n coin flips”) = 1 2 n , and EXAMPLE S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }. ∞ n=1 1 2 n = 1 2 + 1 4 + 1 8 + 1 16 . . .
  • 21. Probability The sample space is not necessarily finite. Flip a coin until first tail shows up: Pr(“It takes n coin flips”) = 1 2 n , and EXAMPLE S = {T, HT, HHT, HHHT, HHHHT, HHHHHT, . . . }. ∞ n=1 1 2 n = 1 2 + 1 4 + 1 8 + 1 16 . . . = 1
  • 22. Event An event is a subset V of the sample space S.
  • 23. Event An event is a subset V of the sample space S. The probability of event V happening, denoted Pr(V ), is Pr(V ) = x∈V Pr(x).
  • 24. Event An event is a subset V of the sample space S. EXAMPLE The probability of event V happening, denoted Pr(V ), is Pr(V ) = x∈V Pr(x). Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH} For each x ∈ S, Pr(x) = 1 8
  • 25. Event An event is a subset V of the sample space S. EXAMPLE The probability of event V happening, denoted Pr(V ), is Pr(V ) = x∈V Pr(x). Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH} For each x ∈ S, Pr(x) = 1 8 Define V to be the event “the first and last coin flips are the same”
  • 26. Event An event is a subset V of the sample space S. EXAMPLE The probability of event V happening, denoted Pr(V ), is Pr(V ) = x∈V Pr(x). Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH} For each x ∈ S, Pr(x) = 1 8 Define V to be the event “the first and last coin flips are the same” in other words, V = {HHH, HTH, THT, TTT}
  • 27. Event An event is a subset V of the sample space S. EXAMPLE The probability of event V happening, denoted Pr(V ), is Pr(V ) = x∈V Pr(x). Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH} For each x ∈ S, Pr(x) = 1 8 Define V to be the event “the first and last coin flips are the same” in other words, V = {HHH, HTH, THT, TTT} What is Pr(V )?
  • 28. Event An event is a subset V of the sample space S. Pr(V ) = Pr(HHH) + Pr(HTH) + Pr(THT) + Pr(TTT) = 4 × 1 8 = 1 2 . EXAMPLE The probability of event V happening, denoted Pr(V ), is Pr(V ) = x∈V Pr(x). Flip a coin 3 times: S = {TTT, TTH, THT, HTT, HHT, HTH, THH, HHH} For each x ∈ S, Pr(x) = 1 8 Define V to be the event “the first and last coin flips are the same” in other words, V = {HHH, HTH, THT, TTT} What is Pr(V )?
  • 29. Random variable A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x).
  • 30. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x).
  • 31. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x).
  • 32. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x). sum over all values of x such that Y (x) = y
  • 33. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x). sum over all values of x such that Y (x) = y What is Pr(Y = 2)?
  • 34. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x). sum over all values of x such that Y (x) = y What is Pr(Y = 2)? Pr(Y = 2) = x∈{HH,TT} Pr(x) = 1 4 + 1 4 = 1 2
  • 35. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x). sum over all values of x such that Y (x) = y What is Pr(Y = 2)? Pr(Y = 2) = x∈{HH,TT} Pr(x) = 1 4 + 1 4 = 1 2
  • 36. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x). sum over all values of x such that Y (x) = y What is Pr(Y = 2)? Pr(Y = 2) = x∈{HH,TT} Pr(x) = 1 4 + 1 4 = 1 2
  • 37. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y Pr(Y = 2) = 1 2 EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x). sum over all values of x such that Y (x) = y What is Pr(Y = 2)? Pr(Y = 2) = x∈{HH,TT} Pr(x) = 1 4 + 1 4 = 1 2
  • 38. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y Pr(Y = 2) = 1 2 EXAMPLE A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x).
  • 39. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y Pr(Y = 2) = 1 2 The expected value (the mean) of a r.v. Y , EXAMPLE denoted E(Y ), is E(Y ) = x∈S Y (x)·Pr(x). A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x).
  • 40. Random variable The probability of Y taking value y is {x ∈ S st. Y(x) = y} Two coin flips. H H H T T H T T 2 1 5 2 S Y Pr(Y = 2) = 1 2 The expected value (the mean) of a r.v. Y , E(Y ) = 2 · 1 2 + 1 · 1 4 + 5 · 1 4 = 7 2 EXAMPLE denoted E(Y ), is E(Y ) = x∈S Y (x)·Pr(x). A random variable (r.v.) Y over sample space S is a function S → R i.e. it maps each outcome x ∈ S to some real number Y (x). Pr(Y = y) = Pr(x).
  • 41. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) THEOREM (Linearity of expectation)
  • 42. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) (regardless of whether the random variables are independent or not.)
  • 43. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values.
  • 44. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. random variable
  • 45. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values.
  • 46. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )?
  • 47. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 1: (without the theorem)
  • 48. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 1: (without the theorem) The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
  • 49. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, E(Y ) = x∈S Y (x) · Pr(x) = 1 36 x∈S Y (x) = THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 1: (without the theorem) The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
  • 50. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, E(Y ) = x∈S Y (x) · Pr(x) = 1 36 x∈S Y (x) = THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 1: (without the theorem) The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes)
  • 51. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, E(Y ) = x∈S Y (x) · Pr(x) = 1 36 x∈S Y (x) = THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 1: (without the theorem) The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes) 1 36 (1 · 2 + 2 · 3 + 3 · 4 + · · · + 1 · 12) = 7
  • 52. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, E(Y ) = x∈S Y (x) · Pr(x) = 1 36 x∈S Y (x) = THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 1: (without the theorem) The sample space S = {(1, 1), (1, 2), (1, 3) . . . (6, 6)} (36 outcomes) 1 36 (1 · 2 + 2 · 3 + 3 · 4 + · · · + 1 · 12) = 7
  • 53. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )?
  • 54. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 2: (with the theorem)
  • 55. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 2: (with the theorem) Let the r.v. Y1 be the value of the first die and Y2 the value of the second
  • 56. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 2: (with the theorem) Let the r.v. Y1 be the value of the first die and Y2 the value of the second E(Y1) = E(Y2) = 3.5
  • 57. Linearity of expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi) Linearity of expectation always holds, THEOREM (Linearity of expectation) EXAMPLE (regardless of whether the random variables are independent or not.) Roll two dice. Let the r.v. Y be the sum of the values. What is E(Y )? Approach 2: (with the theorem) Let the r.v. Y1 be the value of the first die and Y2 the value of the second E(Y1) = E(Y2) = 3.5 so E(Y ) = E(Y1 + Y2) = E(Y1) + E(Y2) = 7
  • 58. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. (usually referred to by the letter I)
  • 59. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. (usually referred to by the letter I) Fact: E(I) = 0 · Pr(I = 0) + 1 · Pr(I = 1) = Pr(I = 1).
  • 60. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. (usually referred to by the letter I) Fact: E(I) = Pr(I = 1).
  • 61. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Often an indicator r.v. I is associated with an event such that (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise).
  • 62. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise).
  • 63. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise).
  • 64. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll?
  • 65. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise)
  • 66. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Pr(Ij = 1) = 21 36 = 7 12 . (by counting the outcomes) Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise)
  • 67. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Pr(Ij = 1) = 21 36 = 7 12 . (by counting the outcomes) Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise) E n j=2 Ij = n j=2 E(Ij) = n j=2 Pr(Ij = 1) = (n − 1) · 7 12
  • 68. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Pr(Ij = 1) = 21 36 = 7 12 . (by counting the outcomes) Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise) E n j=2 Ij = n j=2 E(Ij) = n j=2 Pr(Ij = 1) = (n − 1) · 7 12 Linearity of Expectation Let Y1, Y2, . . . , Yk be k random variables. Then E k i=1 Yi = k i=1 E(Yi)
  • 69. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Pr(Ij = 1) = 21 36 = 7 12 . (by counting the outcomes) Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise) E n j=2 Ij = n j=2 E(Ij) = n j=2 Pr(Ij = 1) = (n − 1) · 7 12
  • 70. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Pr(Ij = 1) = 21 36 = 7 12 . (by counting the outcomes) Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise) E n j=2 Ij = n j=2 E(Ij) = n j=2 Pr(Ij = 1) = (n − 1) · 7 12
  • 71. Indicator random variables An indicator random variable is a r.v. that can only be 0 or 1. Roll a die n times. Pr(Ij = 1) = 21 36 = 7 12 . (by counting the outcomes) Often an indicator r.v. I is associated with an event such that Indicator random variables and linearity of expectation work great together! EXAMPLE (usually referred to by the letter I) Fact: E(I) = Pr(I = 1). I = 1 if the event happens (and I = 0 otherwise). How many rolls do we expect to show a value that is at least the value of the previous roll? For j ∈ {2, . . . , n}, let indicator r.v. Ij = 1 if the value of the jth roll is at least the value of the previous roll (and Ij = 0 otherwise) E n j=2 Ij = n j=2 E(Ij) = n j=2 Pr(Ij = 1) = (n − 1) · 7 12
  • 72. Markov’s inequality EXAMPLE Suppose that the average (mean) speed on the motorway is 60 mph.
  • 73. Markov’s inequality It then follows that at most EXAMPLE Suppose that the average (mean) speed on the motorway is 60 mph.
  • 74. Markov’s inequality It then follows that at most EXAMPLE Suppose that the average (mean) speed on the motorway is 60 mph. 1 2 of all cars drive at least 120 mph,
  • 75. Markov’s inequality It then follows that at most EXAMPLE . . . otherwise the mean must be higher than 60 mph. (a contradiction) Suppose that the average (mean) speed on the motorway is 60 mph. 1 2 of all cars drive at least 120 mph,
  • 76. Markov’s inequality It then follows that at most EXAMPLE . . . otherwise the mean must be higher than 60 mph. (a contradiction) Suppose that the average (mean) speed on the motorway is 60 mph. 2 3 of all cars drive at least 90 mph,
  • 77. Markov’s inequality It then follows that at most If X is a non-negative r.v., then for all a > 0, Pr(X ≥ a) ≤ E(X) a . THEOREM (Markov’s inequality) EXAMPLE . . . otherwise the mean must be higher than 60 mph. (a contradiction) Suppose that the average (mean) speed on the motorway is 60 mph. 2 3 of all cars drive at least 90 mph,
  • 78. Markov’s inequality It then follows that at most If X is a non-negative r.v., then for all a > 0, Pr(X ≥ a) ≤ E(X) a . From the example above: Pr(speed of a random car ≥ 120 mph) ≤ 60 120 = 1 2 , Pr(speed of a random car ≥ 90mph) ≤ 60 90 = 2 3 . EXAMPLE THEOREM (Markov’s inequality) EXAMPLE . . . otherwise the mean must be higher than 60 mph. (a contradiction) Suppose that the average (mean) speed on the motorway is 60 mph. 2 3 of all cars drive at least 90 mph,
  • 79. Markov’s inequality EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat.
  • 80. Markov’s inequality EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat?
  • 81. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? otherwise Ij = 0.
  • 82. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . .
  • 83. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . . Fact: E(I) = Pr(I = 1).
  • 84. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . .
  • 85. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, By Markov’s inequality (recall: Pr(X ≥ a) ≤ E(X) a ), EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . .
  • 86. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, By Markov’s inequality (recall: Pr(X ≥ a) ≤ E(X) a ), EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . . Pr(5 or more people leaving with their own hats) ≤ 1 5 ,
  • 87. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, By Markov’s inequality (recall: Pr(X ≥ a) ≤ E(X) a ), EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . . Pr(5 or more people leaving with their own hats) ≤ 1 5 , Pr(at least 1 person leaving with their own hat) ≤ 1 1 = 1.
  • 88. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, By Markov’s inequality (recall: Pr(X ≥ a) ≤ E(X) a ), (sometimes Markov’s inequality is not particularly informative) EXAMPLE n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . . Pr(5 or more people leaving with their own hats) ≤ 1 5 , Pr(at least 1 person leaving with their own hat) ≤ 1 1 = 1.
  • 89. Markov’s inequality For j ∈ {1, . . . , n}, let indicator r.v. Ij = 1 if the jth person gets their own hat, By Markov’s inequality (recall: Pr(X ≥ a) ≤ E(X) a ), (sometimes Markov’s inequality is not particularly informative) EXAMPLE In fact, here it can be shown that as n → ∞, the probability that at least one person leaves with their own hat is 1 − 1 e . n people go to a party, leaving their hats at the door. Each person leaves with a random hat. How many people leave with their own hat? E n j=1 Ij = n j=1 E(Ij) = n j=1 Pr(Ij = 1) = n· 1 n = 1. otherwise Ij = 0. By linearity of expectation. . . Pr(5 or more people leaving with their own hats) ≤ 1 5 , Pr(at least 1 person leaving with their own hat) ≤ 1 1 = 1.
  • 90. Markov’s inequality If X is a non-negative r.v. that only takes integer values, then Pr(X > 0) = Pr(X ≥ 1) ≤ E(X) . COROLLARY For an indicator r.v. I, the bound is tight (=), as Pr(I > 0) = E(I).
  • 91. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). THEOREM (union bound)
  • 92. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). THEOREM (union bound) This is the probability at least one of the events happens
  • 93. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). THEOREM (union bound)
  • 94. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty)
  • 95. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty)
  • 96. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty)
  • 97. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty) Let the r.v. X = k j=1 Ij be the number of events that happen.
  • 98. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty) Pr k j=1 Vj = Pr(X >0) ≤ E(X) = E( k j=1 Ij) = k j=1 E(Ij) Let the r.v. X = k j=1 Ij be the number of events that happen. = k j=1 Pr(Vj)
  • 99. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. by previous Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty) Pr k j=1 Vj = Pr(X >0) ≤ E(X) = E( k j=1 Ij) = k j=1 E(Ij) Let the r.v. X = k j=1 Ij be the number of events that happen. = k j=1 Pr(Vj) Markov corollary
  • 100. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. by previous Linearity of expectation Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty) Pr k j=1 Vj = Pr(X >0) ≤ E(X) = E( k j=1 Ij) = k j=1 E(Ij) Let the r.v. X = k j=1 Ij be the number of events that happen. = k j=1 Pr(Vj) Markov corollary
  • 101. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. by previous Linearity of expectation Define indicator r.v. Ij to be 1 if event Vj happens, otherwise Ij = 0. THEOREM (union bound) PROOF (Vi and Vj are disjoint iff Vi ∩ Vj is empty) Pr k j=1 Vj = Pr(X >0) ≤ E(X) = E( k j=1 Ij) = k j=1 E(Ij) Let the r.v. X = k j=1 Ij be the number of events that happen. = k j=1 Pr(Vj) Markov corollary E(I) = Pr(I = 1)
  • 102. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty)
  • 103. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. EXAMPLE
  • 104. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. EXAMPLE We define two events: V1 = {3, 4} V2 = {1, 2, 3}
  • 105. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1 3 + 1 2 = 5 6 EXAMPLE We define two events: V1 = {3, 4} V2 = {1, 2, 3}
  • 106. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. 1 2 S V1 V2 4 6 3 5 Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1 3 + 1 2 = 5 6 EXAMPLE We define two events: V1 = {3, 4} V2 = {1, 2, 3}
  • 107. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. 1 2 S V1 V2 4 6 3 5 Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1 3 + 1 2 = 5 6 EXAMPLE We define two events: V1 = {3, 4} V2 = {1, 2, 3} in fact, Pr(V1 ∪ V2) = 2 3
  • 108. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. 1 2 S V1 V2 4 6 3 5 Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1 3 + 1 2 = 5 6 EXAMPLE We define two events: V1 = {3, 4} V2 = {1, 2, 3} in fact, Pr(V1 ∪ V2) = 2 3 (3 was ‘double counted’)
  • 109. Union bound Let V1, . . . , Vk be k events. Then Pr k i=1 Vi ≤ k i=1 Pr(Vi). This bound is tight (=) when the events are all disjoint. THEOREM (union bound) (Vi and Vj are disjoint iff Vi ∩ Vj is empty) S = {1, . . . , 6} is the set of outcomes of a die roll. 1 2 S V1 V2 4 6 3 5 Pr(V1 ∪ V2) ≤ Pr(V1) + Pr(V2) = 1 3 + 1 2 = 5 6 EXAMPLE We define two events: V1 = {3, 4} V2 = {1, 2, 3} in fact, Pr(V1 ∪ V2) = 2 3 (3 was ‘double counted’) Typically the union bound is used when each Pr(Vi) is much smaller than k.
  • 110. Summary The sample space S is the set of outcomes of an experiment. For x ∈ S, the probability of x, written Pr(x), such that x∈S Pr(x) = 1. is a real number between 0 and 1, An event is a subset V of the sample space S, Pr(V ) = x∈V Pr(x) The probability of Y taking value y is {x ∈ S st. Y(x) = y} A random variable (r.v.) Y is a function which maps x ∈ S to S(x) ∈ R Pr(Y = y) = Pr(x). The expected value (the mean) of Y is E(Y ) = x∈S Y (x)·Pr(x). An indicator random variable is a r.v. that can only be 0 or 1. Fact: E(I) = Pr(I = 1). Let V1, . . . , Vk be k events then, THEOREM (union bound) Pr k i=1 Vi ≤ k i=1 Pr(Vi). If X is a non-negative r.v., then for all a > 0, THEOREM (Markov’s inequality) Pr(X ≥ a) ≤ E(X) a . Let Y1,Y2,...,Yk be k random variables then, E k i=1 Yi = k i=1 E(Yi) THEOREM (Linearity of expectation)