SlideShare a Scribd company logo
Some continuous and discrete distributions
Table of contents
I. Continuous distributions and transformation rules.
A. Standard uniform distribution U[0, 1].
B. Uniform distribution U[a, b].
C. Standard normal distribution N(0, 1).
D. Normal distribution N(µ, σ). E. Standard exponential distribution.
F. Exponential distribution with mean λ.
G. Standard Gamma distribution Γ(r, 1).
H. Gamma distribution Γ(r, λ).
II. Discrete distributions and transformation rules.
A. Bernoulli random variables.
B. Binomial distribution.
C. Poisson distribution.
D. Geometric distribution.
E. Negative binomial distribution.
F. Hypergeometric distribution.
1 Continuous distributions.
Each continuous distribution has a “standard” version and a more general
rescaled version. The transformation from one to the other is always of the
form Y = aX + b, with a > 0, and the resulting identities:
fY (y) =
fX
y−b
a
a
(1)
FY (y) = FX
y − b
a
(2)
E(Y ) = aE(X) + b (3)
V ar(Y ) = a2
V ar(X) (4)
MY (t) = ebt
MX(at) (5)
1
1.1 Standard uniform U[0, 1]
This distribution is “pick a random number between 0 and 1”.
fX(x) =
1 if 0 < x < 1
0 otherwise
FX(x) =



0 if x ≤ 0
x if 0 ≤ x ≤ 1
1 if x ≥ 1
E(X) = 1/2
V ar(X) = 1/12
MX(t) =
et
− 1
t
1.2 Uniform U[a, b]
This distribution is “pick a random number between a and b”. To get a
random number between a and b, take a random number between 0 and 1,
multiply it by b − a, and add a. The properties of this random variable are
obtained by applying rules (1–5) to the previous subsection.
fX(x) = 1/(b − a) if a < x < b
0 otherwise
FX(x) =



0 if x ≤ a
x−a
b−a
if a ≤ x ≤ b
1 if x ≥ b
E(X) = (a + b)/2
V ar(X) = (b − a)2
/12
MX(t) =
ebt
− eat
t(b − a)
1.3 Standard normal N(0, 1)
This is the most important distribution in all of probability because of the
Central Limit Theorem, which states that the sums (or averages) of a large
number of independent random variables is approximately normal, no mat-
ter what the original distributions look like. Specifically, if X is a random
variable with mean µ and standard deviation σ, and if X1, X2, . . . are inde-
pendent copies of X, and if Sn = X1 + · · · + Xn, then for large values of n,
2
Sn is approximately normal with mean nµ and standard deviation σ
√
n, and
Sn−nµ√
nσ
is well approximated by the standard normal distribution.
fX(x) = e−x2/2
/
√
2π
FX(x) is given in the table at the back of the book.
E(X) = 0
V ar(X) = 1
MX(t) = et2/2
1.4 Normal N(µ, σ)
To work with a normal random variable X, convert everything to “Z-scores”,
where Z = (X − µ)/σ. Z is then described by the standard normal distribu-
tion, which you can look up in the back of the book. Here are the formulas
for X.
fX(x) = e−(x−µ)2/2σ2
/
√
2πσ2
FX(x) is computed from Z-scores.
E(X) = µ
V ar(X) = σ2
MX(t) = eµt
eσ2t2/2
1.5 Standard exponential
The exponential distribution describes the time beween successive events in
a Poisson process. How long until the next click on my Geiger counter?
How long until this lightbulb burns out? How long until the next campaign
contribution comes in? A key feature is that it is memoryless: a one-year-
old lightbulb has the same change of burning out tomorrow as a brand new
lightbulb.
fX(x) = e−x
, x > 0
FX(x) = 1 − e−x
, x > 0
3
E(X) = 1
V ar(X) = 1
MX(t) = 1/(1 − t)
1.6 Exponential with mean λ
This is obtained by multiplying a standard exponential by λ. Unfortunately,
the letter λ is used differently for Poisson and exponential distributions. If
a Poisson distribution has an average rate of r, then the waiting time is
exponential with mean 1/r. When talking about the Poisson distribution
we’d be inclined to say “λ = rt”, while when talking about the exponential
distribution we’d be inclined to say λ = 1/r.
fX(x) = λ−1
e−x/λ
, x > 0
FX(x) = 1 − e−x/λ
, x > 0
E(X) = λ
V ar(X) = λ2
MX(t) = 1/(1 − λt)
1.7 Standard Gamma distribution
The sum of r independent (standard) exponential random variables is called
a Gamma random variable. It describes the time you need to wait for r
Poisson events to happen (e.g., the time it takes for 10 light bulbs to burn
out, for the Geiger counter to record 10 clicks, or for 10 people to send in
campaign contributions.) The formula for fX isn’t obvious, and that for FX
is complicated, but the others are directly related to those of the exponential
distribution.
fX(x) = xr−1
e−x
/(r − 1)!, x > 0
FX(x) = complicated,
E(X) = r
V ar(X) = r
MX(t) = (1 − t)−r
4
1.8 Gamma distribution Γ(r, λ)
This is a standard Gamma variable multiplied by λ, or equivalently the sum
of r independent exponential variables, each with mean λ
fX(x) = λ−r
xr−1
e−x/λ
/(r − 1)!, x > 0
FX(x) = complicated,
E(X) = λr
V ar(X) = λ2
r
MX(t) = (1 − λt)−r
2 Discrete distributions and transformation
rules.
The discrete random variables we will consider always take on integer values,
so we never rescale them. Also, the cdf FX(x) is rarely useful, with the
notable exception of the geometric distribution. The transformations that
are more relevant are those for adding two independent random variables. If
Z = X + Y , with X and Y independent, then
fZ(z) =
x
fX(x)fY (z − x) (6)
E(Z) = E(X) + E(Y ) (7)
V ar(Z) = V ar(X) + V ar(Y ) (8)
MZ(t) = MX(t)MY (t) (9)
2.1 Bernoulli
A Bernoulli random variable is a variable that can only take on the values 0
and 1. We let p be the probability of 1, and 1 − p the probability of 0. This
example is easy to analyze, and MANY interesting random variables can be
built from this simple building block.
fX(x) =



1 − p if x = 0
p if x = 1
0 otherwise.
5
E(X) = p
V ar(X) = p(1 − p)
MX(t) = 1 + p(et
− 1)
2.2 Binomial
A binomial random variable is the sum of n independent Bernoulli random
variables, all with the same value of p. The usual application is counting
the number of successes in n independent tries at something, where each try
has a probability p of success. It also applies to sampling with replacement,
and is a very good approximation to sampling (without replacement) from
very large populations. When np and n(1 − p) are both large (say, 20 or
bigger), then the binomial distribution is well approximated by the normal
distribution. When n is large (at least 30) and p is small (less than 0.1), then
the binomial distribution is approximated well by the Poisson distribution.
fX(x) =
n
x
px
(1 − p)n−x
E(X) = np
V ar(X) = np(1 − p)
MX(t) = (1 + p(et
− 1))n
2.3 Poisson
The Poisson distribution (pronounced pwah-SON) is the limit of binomial
when n is large and p is small. The correspondence is λ = np. The Poisson
distribution replicates itself, in that the sum of a Poisson(λ) random variable
and a (independent!) Poisson(µ) random variable is a Poisson(λ+µ) random
variable. Anything that involves the sum of many, many long-shot events
(e.g., number of people hit by lightning in a year, or number of broken bones
from playing soccer, or number of clicks on a Geiger counter) will be described
by the Poisson distribution.
Closely related to the Poisson distribution is a Poisson process. In a Pois-
son process events accumulate at a certain average rate r. The number of
events in “time” t is then given by Poisson(rt). Examples include radioac-
tivity, where you have clicks per unit time, typos (where r is errors/page and
6
t is the number of pages), industrial defects (r equals the average number of
defects per foot of sheet metal and t is the number of feet).
fX(x) = λx
e−λ
/x!
E(X) = λ
V ar(X) = λ
MX(t) = eλ(et−1)
2.4 Geometric
The geometric distribution describes the waiting time between successes in
a Binomial process. For instance, flip coins and count the turns until the
first head, or roll dice and count the turns until you get a “6”. It is very
much like the exponential distribution, with λ corresponding to 1/p, except
that the geometric distribution is discrete while the exponential distribution
is continuous.
fX(x) = pqx−1
, x = 1, 2, . . . , where q = 1 − p
E(X) = 1/p
V ar(X) = q/p2
MX(t) =
pet
1 − qet
2.5 Negative binomial
The sum X of r independent geometric random variables is given by the
discrete analog of the Gamma distribution (which describes the sum of r
independent exponential random variables).
fX(x) =
x − 1
r − 1
pr
qx−r
, x = r, r + 1, . . . , where q = 1 − p
E(X) = r/p
V ar(X) = rq/p2
MX(t) =
pr
ert
(1 − qet)r
7
The term “negative binomial distribution” actually refers to Y = X − r,
and not to X. The data for Y are easily obtained from those of X:
fY (n) =
n + r − 1
n
pr
qn
, n = 1, 2, . . . , where q = 1 − p
E(Y ) = rq/p
V ar(Y ) = rq/p2
MY (t) =
pr
(1 − qet)r
2.6 Hypergeometric distribution
The hypergeometric distribution describes sampling without replacement.
There are r red and w white balls in an urn (with N = r + w), we draw
n balls out, and let X be the number of red balls drawn. Of course, it
doesn’t have to be balls in urns. We could just as well be counting aces
dealt from a standard deck (with 4 aces and 48 non-aces), or Libertarians
drawn from a sample of voters. A hypergeometric random variable is the
sum of n Bernoulli random variables (with Xi indicating the color of the
i-th ball drawn). These Bernoulli variables are NOT independent, but their
covariances are easy to compute, and this gives us the variance of the hyper-
geometric variable. When n N, the hypergeometric distribution is very
close to a Binomial distribution with p = r/N.
fX(x) =
r
x
w
n−x
N
n
E(Y ) = nr/N
V ar(Y ) = n
r
N
w
N
N − n
N − 1
8

More Related Content

What's hot

What's hot (20)

Mit18 330 s12_chapter4
Mit18 330 s12_chapter4Mit18 330 s12_chapter4
Mit18 330 s12_chapter4
 
Fourier series of odd functions with period 2 l
Fourier series of odd functions with period 2 lFourier series of odd functions with period 2 l
Fourier series of odd functions with period 2 l
 
160511 hasegawa lab_seminar
160511 hasegawa lab_seminar160511 hasegawa lab_seminar
160511 hasegawa lab_seminar
 
"reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli..."reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli...
 
Nokton theory-en
Nokton theory-enNokton theory-en
Nokton theory-en
 
Partial differentiation
Partial differentiationPartial differentiation
Partial differentiation
 
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALESNONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
 
Multivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributionsMultivariate Gaussin, Rayleigh & Rician distributions
Multivariate Gaussin, Rayleigh & Rician distributions
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Comparison Theorems for SDEs
Comparison Theorems for SDEs Comparison Theorems for SDEs
Comparison Theorems for SDEs
 
BlUP and BLUE- REML of linear mixed model
BlUP and BLUE- REML of linear mixed modelBlUP and BLUE- REML of linear mixed model
BlUP and BLUE- REML of linear mixed model
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”
 
So a webinar-2013-2
So a webinar-2013-2So a webinar-2013-2
So a webinar-2013-2
 
Survival analysis 1
Survival analysis 1Survival analysis 1
Survival analysis 1
 
Introduction to Decision Making Theory
Introduction to Decision Making TheoryIntroduction to Decision Making Theory
Introduction to Decision Making Theory
 
Partial Differentiation & Application
Partial Differentiation & Application Partial Differentiation & Application
Partial Differentiation & Application
 
1 - Linear Regression
1 - Linear Regression1 - Linear Regression
1 - Linear Regression
 
A short remark on Feller’s square root condition.
A short remark on Feller’s square root condition.A short remark on Feller’s square root condition.
A short remark on Feller’s square root condition.
 
Big model, big data
Big model, big dataBig model, big data
Big model, big data
 
Stochastic Processes - part 6
Stochastic Processes - part 6Stochastic Processes - part 6
Stochastic Processes - part 6
 

Similar to Distributions

Chapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability DistributionsChapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability Distributions
JasonTagapanGulla
 

Similar to Distributions (20)

Doe02 statistics
Doe02 statisticsDoe02 statistics
Doe02 statistics
 
multivariate normal distribution.pdf
multivariate normal distribution.pdfmultivariate normal distribution.pdf
multivariate normal distribution.pdf
 
Multivariate Distributions, an overview
Multivariate Distributions, an overviewMultivariate Distributions, an overview
Multivariate Distributions, an overview
 
Econometrics 2.pptx
Econometrics 2.pptxEconometrics 2.pptx
Econometrics 2.pptx
 
Numarical values
Numarical valuesNumarical values
Numarical values
 
Numarical values highlighted
Numarical values highlightedNumarical values highlighted
Numarical values highlighted
 
Distributionworkshop 2.pptx
Distributionworkshop 2.pptxDistributionworkshop 2.pptx
Distributionworkshop 2.pptx
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
Basics of probability in statistical simulation and stochastic programming
Basics of probability in statistical simulation and stochastic programmingBasics of probability in statistical simulation and stochastic programming
Basics of probability in statistical simulation and stochastic programming
 
The Wishart and inverse-wishart distribution
 The Wishart and inverse-wishart distribution The Wishart and inverse-wishart distribution
The Wishart and inverse-wishart distribution
 
Chapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability DistributionsChapter 3 – Random Variables and Probability Distributions
Chapter 3 – Random Variables and Probability Distributions
 
Generating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceGenerating Chebychev Chaotic Sequence
Generating Chebychev Chaotic Sequence
 
stochastic notes
stochastic notes stochastic notes
stochastic notes
 
Slides mc gill-v4
Slides mc gill-v4Slides mc gill-v4
Slides mc gill-v4
 
Quantitative Techniques random variables
Quantitative Techniques random variablesQuantitative Techniques random variables
Quantitative Techniques random variables
 
Section3 stochastic
Section3 stochasticSection3 stochastic
Section3 stochastic
 
Slides mc gill-v3
Slides mc gill-v3Slides mc gill-v3
Slides mc gill-v3
 
Common Fixed Theorems Using Random Implicit Iterative Schemes
Common Fixed Theorems Using Random Implicit Iterative SchemesCommon Fixed Theorems Using Random Implicit Iterative Schemes
Common Fixed Theorems Using Random Implicit Iterative Schemes
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)
 
The Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability DistributionThe Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability Distribution
 

Recently uploaded

Additional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdfAdditional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdf
joachimlavalley1
 
Accounting and finance exit exam 2016 E.C.pdf
Accounting and finance exit exam 2016 E.C.pdfAccounting and finance exit exam 2016 E.C.pdf
Accounting and finance exit exam 2016 E.C.pdf
YibeltalNibretu
 
Industrial Training Report- AKTU Industrial Training Report
Industrial Training Report- AKTU Industrial Training ReportIndustrial Training Report- AKTU Industrial Training Report
Industrial Training Report- AKTU Industrial Training Report
Avinash Rai
 

Recently uploaded (20)

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
 
Students, digital devices and success - Andreas Schleicher - 27 May 2024..pptx
Students, digital devices and success - Andreas Schleicher - 27 May 2024..pptxStudents, digital devices and success - Andreas Schleicher - 27 May 2024..pptx
Students, digital devices and success - Andreas Schleicher - 27 May 2024..pptx
 
Fish and Chips - have they had their chips
Fish and Chips - have they had their chipsFish and Chips - have they had their chips
Fish and Chips - have they had their chips
 
Additional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdfAdditional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdf
 
Salient features of Environment protection Act 1986.pptx
Salient features of Environment protection Act 1986.pptxSalient features of Environment protection Act 1986.pptx
Salient features of Environment protection Act 1986.pptx
 
UNIT – IV_PCI Complaints: Complaints and evaluation of complaints, Handling o...
UNIT – IV_PCI Complaints: Complaints and evaluation of complaints, Handling o...UNIT – IV_PCI Complaints: Complaints and evaluation of complaints, Handling o...
UNIT – IV_PCI Complaints: Complaints and evaluation of complaints, Handling o...
 
PART A. Introduction to Costumer Service
PART A. Introduction to Costumer ServicePART A. Introduction to Costumer Service
PART A. Introduction to Costumer Service
 
Accounting and finance exit exam 2016 E.C.pdf
Accounting and finance exit exam 2016 E.C.pdfAccounting and finance exit exam 2016 E.C.pdf
Accounting and finance exit exam 2016 E.C.pdf
 
How to Create Map Views in the Odoo 17 ERP
How to Create Map Views in the Odoo 17 ERPHow to Create Map Views in the Odoo 17 ERP
How to Create Map Views in the Odoo 17 ERP
 
INU_CAPSTONEDESIGN_비밀번호486_업로드용 발표자료.pdf
INU_CAPSTONEDESIGN_비밀번호486_업로드용 발표자료.pdfINU_CAPSTONEDESIGN_비밀번호486_업로드용 발표자료.pdf
INU_CAPSTONEDESIGN_비밀번호486_업로드용 발표자료.pdf
 
Instructions for Submissions thorugh G- Classroom.pptx
Instructions for Submissions thorugh G- Classroom.pptxInstructions for Submissions thorugh G- Classroom.pptx
Instructions for Submissions thorugh G- Classroom.pptx
 
Solid waste management & Types of Basic civil Engineering notes by DJ Sir.pptx
Solid waste management & Types of Basic civil Engineering notes by DJ Sir.pptxSolid waste management & Types of Basic civil Engineering notes by DJ Sir.pptx
Solid waste management & Types of Basic civil Engineering notes by DJ Sir.pptx
 
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXXPhrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
 
Benefits and Challenges of Using Open Educational Resources
Benefits and Challenges of Using Open Educational ResourcesBenefits and Challenges of Using Open Educational Resources
Benefits and Challenges of Using Open Educational Resources
 
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
 
Industrial Training Report- AKTU Industrial Training Report
Industrial Training Report- AKTU Industrial Training ReportIndustrial Training Report- AKTU Industrial Training Report
Industrial Training Report- AKTU Industrial Training Report
 
Basic Civil Engg Notes_Chapter-6_Environment Pollution & Engineering
Basic Civil Engg Notes_Chapter-6_Environment Pollution & EngineeringBasic Civil Engg Notes_Chapter-6_Environment Pollution & Engineering
Basic Civil Engg Notes_Chapter-6_Environment Pollution & Engineering
 
Introduction to Quality Improvement Essentials
Introduction to Quality Improvement EssentialsIntroduction to Quality Improvement Essentials
Introduction to Quality Improvement Essentials
 
NLC-2024-Orientation-for-RO-SDO (1).pptx
NLC-2024-Orientation-for-RO-SDO (1).pptxNLC-2024-Orientation-for-RO-SDO (1).pptx
NLC-2024-Orientation-for-RO-SDO (1).pptx
 
Jose-Rizal-and-Philippine-Nationalism-National-Symbol-2.pptx
Jose-Rizal-and-Philippine-Nationalism-National-Symbol-2.pptxJose-Rizal-and-Philippine-Nationalism-National-Symbol-2.pptx
Jose-Rizal-and-Philippine-Nationalism-National-Symbol-2.pptx
 

Distributions

  • 1. Some continuous and discrete distributions Table of contents I. Continuous distributions and transformation rules. A. Standard uniform distribution U[0, 1]. B. Uniform distribution U[a, b]. C. Standard normal distribution N(0, 1). D. Normal distribution N(µ, σ). E. Standard exponential distribution. F. Exponential distribution with mean λ. G. Standard Gamma distribution Γ(r, 1). H. Gamma distribution Γ(r, λ). II. Discrete distributions and transformation rules. A. Bernoulli random variables. B. Binomial distribution. C. Poisson distribution. D. Geometric distribution. E. Negative binomial distribution. F. Hypergeometric distribution. 1 Continuous distributions. Each continuous distribution has a “standard” version and a more general rescaled version. The transformation from one to the other is always of the form Y = aX + b, with a > 0, and the resulting identities: fY (y) = fX y−b a a (1) FY (y) = FX y − b a (2) E(Y ) = aE(X) + b (3) V ar(Y ) = a2 V ar(X) (4) MY (t) = ebt MX(at) (5) 1
  • 2. 1.1 Standard uniform U[0, 1] This distribution is “pick a random number between 0 and 1”. fX(x) = 1 if 0 < x < 1 0 otherwise FX(x) =    0 if x ≤ 0 x if 0 ≤ x ≤ 1 1 if x ≥ 1 E(X) = 1/2 V ar(X) = 1/12 MX(t) = et − 1 t 1.2 Uniform U[a, b] This distribution is “pick a random number between a and b”. To get a random number between a and b, take a random number between 0 and 1, multiply it by b − a, and add a. The properties of this random variable are obtained by applying rules (1–5) to the previous subsection. fX(x) = 1/(b − a) if a < x < b 0 otherwise FX(x) =    0 if x ≤ a x−a b−a if a ≤ x ≤ b 1 if x ≥ b E(X) = (a + b)/2 V ar(X) = (b − a)2 /12 MX(t) = ebt − eat t(b − a) 1.3 Standard normal N(0, 1) This is the most important distribution in all of probability because of the Central Limit Theorem, which states that the sums (or averages) of a large number of independent random variables is approximately normal, no mat- ter what the original distributions look like. Specifically, if X is a random variable with mean µ and standard deviation σ, and if X1, X2, . . . are inde- pendent copies of X, and if Sn = X1 + · · · + Xn, then for large values of n, 2
  • 3. Sn is approximately normal with mean nµ and standard deviation σ √ n, and Sn−nµ√ nσ is well approximated by the standard normal distribution. fX(x) = e−x2/2 / √ 2π FX(x) is given in the table at the back of the book. E(X) = 0 V ar(X) = 1 MX(t) = et2/2 1.4 Normal N(µ, σ) To work with a normal random variable X, convert everything to “Z-scores”, where Z = (X − µ)/σ. Z is then described by the standard normal distribu- tion, which you can look up in the back of the book. Here are the formulas for X. fX(x) = e−(x−µ)2/2σ2 / √ 2πσ2 FX(x) is computed from Z-scores. E(X) = µ V ar(X) = σ2 MX(t) = eµt eσ2t2/2 1.5 Standard exponential The exponential distribution describes the time beween successive events in a Poisson process. How long until the next click on my Geiger counter? How long until this lightbulb burns out? How long until the next campaign contribution comes in? A key feature is that it is memoryless: a one-year- old lightbulb has the same change of burning out tomorrow as a brand new lightbulb. fX(x) = e−x , x > 0 FX(x) = 1 − e−x , x > 0 3
  • 4. E(X) = 1 V ar(X) = 1 MX(t) = 1/(1 − t) 1.6 Exponential with mean λ This is obtained by multiplying a standard exponential by λ. Unfortunately, the letter λ is used differently for Poisson and exponential distributions. If a Poisson distribution has an average rate of r, then the waiting time is exponential with mean 1/r. When talking about the Poisson distribution we’d be inclined to say “λ = rt”, while when talking about the exponential distribution we’d be inclined to say λ = 1/r. fX(x) = λ−1 e−x/λ , x > 0 FX(x) = 1 − e−x/λ , x > 0 E(X) = λ V ar(X) = λ2 MX(t) = 1/(1 − λt) 1.7 Standard Gamma distribution The sum of r independent (standard) exponential random variables is called a Gamma random variable. It describes the time you need to wait for r Poisson events to happen (e.g., the time it takes for 10 light bulbs to burn out, for the Geiger counter to record 10 clicks, or for 10 people to send in campaign contributions.) The formula for fX isn’t obvious, and that for FX is complicated, but the others are directly related to those of the exponential distribution. fX(x) = xr−1 e−x /(r − 1)!, x > 0 FX(x) = complicated, E(X) = r V ar(X) = r MX(t) = (1 − t)−r 4
  • 5. 1.8 Gamma distribution Γ(r, λ) This is a standard Gamma variable multiplied by λ, or equivalently the sum of r independent exponential variables, each with mean λ fX(x) = λ−r xr−1 e−x/λ /(r − 1)!, x > 0 FX(x) = complicated, E(X) = λr V ar(X) = λ2 r MX(t) = (1 − λt)−r 2 Discrete distributions and transformation rules. The discrete random variables we will consider always take on integer values, so we never rescale them. Also, the cdf FX(x) is rarely useful, with the notable exception of the geometric distribution. The transformations that are more relevant are those for adding two independent random variables. If Z = X + Y , with X and Y independent, then fZ(z) = x fX(x)fY (z − x) (6) E(Z) = E(X) + E(Y ) (7) V ar(Z) = V ar(X) + V ar(Y ) (8) MZ(t) = MX(t)MY (t) (9) 2.1 Bernoulli A Bernoulli random variable is a variable that can only take on the values 0 and 1. We let p be the probability of 1, and 1 − p the probability of 0. This example is easy to analyze, and MANY interesting random variables can be built from this simple building block. fX(x) =    1 − p if x = 0 p if x = 1 0 otherwise. 5
  • 6. E(X) = p V ar(X) = p(1 − p) MX(t) = 1 + p(et − 1) 2.2 Binomial A binomial random variable is the sum of n independent Bernoulli random variables, all with the same value of p. The usual application is counting the number of successes in n independent tries at something, where each try has a probability p of success. It also applies to sampling with replacement, and is a very good approximation to sampling (without replacement) from very large populations. When np and n(1 − p) are both large (say, 20 or bigger), then the binomial distribution is well approximated by the normal distribution. When n is large (at least 30) and p is small (less than 0.1), then the binomial distribution is approximated well by the Poisson distribution. fX(x) = n x px (1 − p)n−x E(X) = np V ar(X) = np(1 − p) MX(t) = (1 + p(et − 1))n 2.3 Poisson The Poisson distribution (pronounced pwah-SON) is the limit of binomial when n is large and p is small. The correspondence is λ = np. The Poisson distribution replicates itself, in that the sum of a Poisson(λ) random variable and a (independent!) Poisson(µ) random variable is a Poisson(λ+µ) random variable. Anything that involves the sum of many, many long-shot events (e.g., number of people hit by lightning in a year, or number of broken bones from playing soccer, or number of clicks on a Geiger counter) will be described by the Poisson distribution. Closely related to the Poisson distribution is a Poisson process. In a Pois- son process events accumulate at a certain average rate r. The number of events in “time” t is then given by Poisson(rt). Examples include radioac- tivity, where you have clicks per unit time, typos (where r is errors/page and 6
  • 7. t is the number of pages), industrial defects (r equals the average number of defects per foot of sheet metal and t is the number of feet). fX(x) = λx e−λ /x! E(X) = λ V ar(X) = λ MX(t) = eλ(et−1) 2.4 Geometric The geometric distribution describes the waiting time between successes in a Binomial process. For instance, flip coins and count the turns until the first head, or roll dice and count the turns until you get a “6”. It is very much like the exponential distribution, with λ corresponding to 1/p, except that the geometric distribution is discrete while the exponential distribution is continuous. fX(x) = pqx−1 , x = 1, 2, . . . , where q = 1 − p E(X) = 1/p V ar(X) = q/p2 MX(t) = pet 1 − qet 2.5 Negative binomial The sum X of r independent geometric random variables is given by the discrete analog of the Gamma distribution (which describes the sum of r independent exponential random variables). fX(x) = x − 1 r − 1 pr qx−r , x = r, r + 1, . . . , where q = 1 − p E(X) = r/p V ar(X) = rq/p2 MX(t) = pr ert (1 − qet)r 7
  • 8. The term “negative binomial distribution” actually refers to Y = X − r, and not to X. The data for Y are easily obtained from those of X: fY (n) = n + r − 1 n pr qn , n = 1, 2, . . . , where q = 1 − p E(Y ) = rq/p V ar(Y ) = rq/p2 MY (t) = pr (1 − qet)r 2.6 Hypergeometric distribution The hypergeometric distribution describes sampling without replacement. There are r red and w white balls in an urn (with N = r + w), we draw n balls out, and let X be the number of red balls drawn. Of course, it doesn’t have to be balls in urns. We could just as well be counting aces dealt from a standard deck (with 4 aces and 48 non-aces), or Libertarians drawn from a sample of voters. A hypergeometric random variable is the sum of n Bernoulli random variables (with Xi indicating the color of the i-th ball drawn). These Bernoulli variables are NOT independent, but their covariances are easy to compute, and this gives us the variance of the hyper- geometric variable. When n N, the hypergeometric distribution is very close to a Binomial distribution with p = r/N. fX(x) = r x w n−x N n E(Y ) = nr/N V ar(Y ) = n r N w N N − n N − 1 8