A random variable may assume any value between 10 and 50 with equal likelihoods. (Uniform
distribution) Determine the following values for this probability distribution:
a) µ
b) s
c) f(x) =
d) P(x < 25)
e) P(x > 15)
f) P( 12 < x < 30)
Solution
The distribution of a random variable X for which the probability density function f
is given by The parameters µ and s2 are, respectively, the mean and variance of the
distribution. The distribution is denoted by N(µ, s2). If the random variable X has such a
distribution, then this is denoted by X ~ N(µ, s2) and the random variable may be referred to as a
normal variable. The graph of f(x) approaches the x-axis extremely quickly, and is effectively
zero if |x-µ| < 3s (hence the three-sigma rule). In fact, P(|X-µ| < 2s)˜95.5% and P(|X-µ| <
3s)˜99.7%. The first derivation of the form of f is believed to be that of de Moivre in 1733. The
description \'normal distribution\' was used by Galton in 1889, whereas \'Gaussian distribution\'
was used by Karl Pearson in 1905. The normal distribution is the basis of a large proportion of
statistical analysis. Its importance and ubiquity are largely a consequence of the Central Limit
Theorem, which implies that averaging almost always leads to a bell-shaped distribution (hence
the name \'normal\'). See bell-curve. Normal distribution. The diagram illustrates the
probability density function of a normal random variable X having expectation µ and variance
s2. The distribution has mean, median, and mode at x=µ, where the density function has value
1/(svp). Note that almost all the distribution (99.7%) lies within 3s of the central value. The
standard normal distribution has mean 0 and variance 1. A random variable with this distribution
is often denoted by Z and we write Z ~ N(0, 1). Its probability density function is usually
denoted by ? and is given by If X has a general normal distribution N(µ, s2) then Z, defined by
the standardizing transformation has a standard normal distribution. It follows that the graph of
the probability density function of X is obtained from the corresponding graph for Z by a stretch
parallel to the z-axis, with centre at the origin and scale-factor s, followed by a translation along
the z-axis by µ. Standard normal distribution. The distribution is centred on 0, with 99.7%
falling between -3 and 3 and 95% falling between -1.96 and 1.96. The cumulative probability
function of Z is usually denoted by F and tables of values of F(z) are commonly available (see
The Standard Normal Distribution Function; see also Upper-Tail Percentage Points for the
Standard Normal Distribution). These tables usually give F(z) only for z>0, since values for
negative values of z can be found using F(z)=1-F(-z).The tables can be used to find cumulative
probabilities for X ~ N(µ, s2) via the standardizing transformation given above, since, for
example, As an example, if X ~ N(7, 25) then the probability of X taking a value between 5
and 10 is given by The normal d.
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
Uniform Random Variables: Calculating Probability Values
1. A random variable may assume any value between 10 and 50 with equal likelihoods. (Uniform
distribution) Determine the following values for this probability distribution:
a) µ
b) s
c) f(x) =
d) P(x < 25)
e) P(x > 15)
f) P( 12 < x < 30)
Solution
The distribution of a random variable X for which the probability density function f
is given by The parameters µ and s2 are, respectively, the mean and variance of the
distribution. The distribution is denoted by N(µ, s2). If the random variable X has such a
distribution, then this is denoted by X ~ N(µ, s2) and the random variable may be referred to as a
normal variable. The graph of f(x) approaches the x-axis extremely quickly, and is effectively
zero if |x-µ| < 3s (hence the three-sigma rule). In fact, P(|X-µ| < 2s)˜95.5% and P(|X-µ| <
3s)˜99.7%. The first derivation of the form of f is believed to be that of de Moivre in 1733. The
description 'normal distribution' was used by Galton in 1889, whereas 'Gaussian distribution'
was used by Karl Pearson in 1905. The normal distribution is the basis of a large proportion of
statistical analysis. Its importance and ubiquity are largely a consequence of the Central Limit
Theorem, which implies that averaging almost always leads to a bell-shaped distribution (hence
the name 'normal'). See bell-curve. Normal distribution. The diagram illustrates the
probability density function of a normal random variable X having expectation µ and variance
s2. The distribution has mean, median, and mode at x=µ, where the density function has value
1/(svp). Note that almost all the distribution (99.7%) lies within 3s of the central value. The
standard normal distribution has mean 0 and variance 1. A random variable with this distribution
is often denoted by Z and we write Z ~ N(0, 1). Its probability density function is usually
2. denoted by ? and is given by If X has a general normal distribution N(µ, s2) then Z, defined by
the standardizing transformation has a standard normal distribution. It follows that the graph of
the probability density function of X is obtained from the corresponding graph for Z by a stretch
parallel to the z-axis, with centre at the origin and scale-factor s, followed by a translation along
the z-axis by µ. Standard normal distribution. The distribution is centred on 0, with 99.7%
falling between -3 and 3 and 95% falling between -1.96 and 1.96. The cumulative probability
function of Z is usually denoted by F and tables of values of F(z) are commonly available (see
The Standard Normal Distribution Function; see also Upper-Tail Percentage Points for the
Standard Normal Distribution). These tables usually give F(z) only for z>0, since values for
negative values of z can be found using F(z)=1-F(-z).The tables can be used to find cumulative
probabilities for X ~ N(µ, s2) via the standardizing transformation given above, since, for
example, As an example, if X ~ N(7, 25) then the probability of X taking a value between 5
and 10 is given by The normal distribution plays a central part in the theory of errors that was
developed by Gauss. In the theory of errors, the error function (erf) is defined by erf(x)=2F(xv2)-
1.An important property of the normal distribution is that any linear combination of independent
normal variables is normal: if are independent, and a and b are constants,
thenaX1+bX2~N(aµ1+bµ2,a2s21+b2s22),with the obvious generalization to n independent
normal variables. Many distributions can be approximated by a normal distribution for suitably
large values of the relevant parameters. See also binomial distribution; chi-squared distribution;
Poisson distribution; t-distribution.