SlideShare a Scribd company logo
Elementary Statistical Inference 1
석 진 환
GROOT SEMINAR
Reference
• Introduction to Mathematical Statistics, Hogg, 7e
Which is better: machine learning or statistics?
• Prediction vs. Explanation
• Machine learning models are designed to make the most accurate
predictions possible.
• Statistical models are designed for inference about the relationships
between variables.
• Many statistical models can make predictions, but predictive accuracy is
not their strength. Likewise, machine learning models provide various
degrees of interpretability
목차
1. Estimation
1) Point estimation
2) Interval estimation
3) Order Statistics
2. Hypothesis Testing
3. The Method of Monte Carlo & Bootstrap Procedure
Sampling and Statistics
• We have a random variable 𝑋 of interest, but its pdf 𝑓 𝑥 or pmf 𝑝(𝑥) is
not known
1. 𝑓 𝑥 or 𝑝(𝑥) is completely unknown
2. The form of 𝑓 𝑥 or 𝑝(𝑥) is know, down to a parameter 𝜃
Our information about the unknown distribution or the unknown
parameters of the distribution of 𝑋 comes from a sample on 𝑋.
1. Sampling and Statistics
• We consider the second classification(parametric inference)
• 𝑋 has an exponential distribution, 𝐸𝑥𝑝 𝜃 , 𝜃 is unknown.
• 𝑋 has a binomial distribution, b 𝑛, 𝑝 where 𝑛 is known, but 𝑝 is
unknown.
• 𝑋 has a gamma distribution, Γ 𝛼, 𝛽 , 𝛼 and 𝛽 are unknown.
• 𝑓 𝑥 ; 𝜃 , 𝑝 𝑥 ; 𝜃 𝜃 ∈ Ω
1. Sampling and Statistics
• The sample observations have the same distribution as 𝑋,
and we denote them as the random variables 𝑋1, 𝑋2, … , 𝑋 𝑛,
where 𝑛 denotes the sample size.
• When the sample is actually drawn, we use lower case letters
𝑥1, 𝑥2, … , 𝑥 𝑛 as realizations of the sample.
• Assume that 𝑋1, 𝑋2, … , 𝑋 𝑛 are mutually independent.
• Definition 1
• If 𝑋1, 𝑋2, … , 𝑋 𝑛 iid, then these random variables constitute a random
sample of size n from the common distribution.
• vs Stochastic Process ??
1. Sampling and Statistics
• Definition 2 (Statistic)
• 𝑋1, 𝑋2, … , 𝑋 𝑛 sample on a random variable 𝑋. Let 𝑇 = 𝑇(𝑋1, 𝑋2, … , 𝑋 𝑛) be a
function of the sample. Then 𝑇 is called a statistic(통계량)
• Once the sample is actually drawn, then 𝑡 is called the realization of 𝑇,
where 𝑡 = 𝑇(𝑥1, 𝑥2, … , 𝑥 𝑛)
• Example : 𝑋 =
𝑋1+𝑋2+⋯+𝑋 𝑛
𝑛
, 𝑆2 =
1
𝑛−1
𝑋𝑖 − 𝑋 2 are statistics
• It makes sense to consider a statistic 𝑇, which is an estimator of 𝜃.
• While we call 𝑇 an estimator of 𝜃, we call its realization 𝑡 an estimate of 𝜃.
1. Sampling and Statistics
Estimation
• Point estimation
• 표본평균이 15니까, 모평균도 15일것이다!
• MLE
• Interval estimation
• 표본평균이 15니까, 모평균은 (12.3, 14.7) 구간에 있을 것이다!
• Confidence Interval
1. Sampling and Statistics
Several Properties of point estimators
• Unbiasedness(비편향성), Consistency(일치성), efficiency(효율성)
• Definition 3 (Unbiasedness)
• 𝑋1, 𝑋2, … , 𝑋 𝑛 : sample on a random variable 𝑋 with pdf 𝑓 𝑥 ; 𝜃 , 𝜃 ∈ Ω.
Let 𝑇 = 𝑇(𝑋1, 𝑋2, … , 𝑋 𝑛) be a statistic. We say that 𝑇 is unbiased
estimator of 𝜃 if 𝐸 𝑇 = 𝜃.
1. Sampling and Statistics
Example : Maximum Likelihood Estimator(MLE)
• 𝐿 𝜃 = 𝑖=1
𝑛
𝑓(𝑥𝑖, 𝜃), likelihood function of the random
sample.
• Often-used estimate is that value of 𝜃 which provides a
maximum of 𝐿(𝜃). If it is unique, this is called the maximum
likelihood estimator, and denote it as 𝜃,
i.e., 𝜃 = 𝐴𝑟𝑔𝑚𝑎𝑥 𝜃 𝐿(𝜃).
1. Sampling and Statistics
MLE with EXPONENTIAL DISTRIBUTION
• Random sample 𝑋1, 𝑋2, … , 𝑋 𝑛 is the Γ(1, 𝜃) density. The log of
likelihood function is
𝑙 𝜃 = 𝑙𝑜𝑔
𝑖=1
𝑛
1
𝜃
𝑒
−
𝑥 𝑖
𝜃 = −𝑛log𝜃 − 𝜃−1
𝑖=1
𝑛
𝑥𝑖
•
𝜕𝑙 𝜃
𝜕𝜃
= 0,
𝜕2 𝑙 𝜃
𝜕𝜃2 < 0 yields that 𝜃 = 𝑋 is the mle of 𝜃.
• Because 𝐸 𝑋 = 𝜃, we have that 𝐸 𝑋 = 𝜃 and, hence, 𝜃 is an
unbiased estimator of 𝜃.
1. Sampling and Statistics
MLE with NORMAL DISTRIBUTION
• 𝑋 ~ 𝑁 𝜇, 𝜎2
∗ 𝜽 = 𝜇, 𝜎
• 𝑓 𝑥; 𝜽 =
1
2𝜋𝜎
𝑒
−
𝑥−𝜇 2
2𝜎2
• If 𝑋1, 𝑋2, … , 𝑋 𝑛 is a random variable on 𝑋, then
𝐿 𝜽 =
𝑖=1
𝑛
𝑓 𝑥𝑖; 𝜽
𝑙 𝜽 = log 𝐿 𝜽 =
𝑖=1
𝑛
log(𝑓(𝑥𝑖; 𝜃)) = −
𝑛
2
log 2𝜋 − 𝑛 log 𝜎 −
1
2
𝑖=1
𝑛
𝑥𝑖 − 𝜇
𝜎
2
𝜕𝑙
𝜕𝜇
=
𝜕𝑙
𝜕𝜎
= 0 yields 𝜇 = 𝑋 𝑎𝑛𝑑 𝜎2 = 𝑛−1
𝑖=1
𝑛
𝑋𝑖 − 𝑋 2 .
• 𝜇 is an unbiased estimator of 𝜇
• 𝜎2 is a biased estimator of 𝜎2 , though converges to 𝜎2 as → ∞.
1. Sampling and Statistics
예제1(점추정)
베스킨라빈스 파인트의 무게는 정규분포를 따른다고 가정한다고 알고
있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다.
MLE에 의한 파인트 무게의 모평균의 추정값은??
표본 평균 : 321g -> 모평균의 추정값이 된다!
만약 회사 홈페이지에 310g이라고 써져있다면?
가설검정
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
2. Confidence Intervals
Confidence Intervals
• We discussed estimating 𝜃 by a statistic 𝜃 = 𝜃 𝑋1, 𝑋2, … , 𝑋 𝑛 ,
where 𝑋1, 𝑋2, … , 𝑋 𝑛 is a sample from the distribution of 𝑋.
• When the sample is drawn, it is unlikely that the value of 𝜃 is
the true value of the parameter.
• Error must exist.
• Embody estimate of error in terms of a confidence interval
2. Confidence Intervals
Confidence Intervals
• Definition 4 (Confidence Interval)
• Let 𝑋1, 𝑋2, … , 𝑋 𝑛 be a sample on a random variable 𝑋, where 𝑋 has pdf
𝑓 𝑥: 𝜃 , 𝜃 ∈ Ω. Let 𝛼 be specified. Let 𝐿 = 𝐿 𝑋1, 𝑋2, … , 𝑋 𝑛 and 𝑈 =
𝑈 𝑋1, 𝑋2, … , 𝑋 𝑛 be two statistics.
• We say that the interval (𝐿, 𝑈) is a 1 − 𝛼 100% confidence interval(신뢰
구간) for 𝜃 if
1 − 𝛼 = 𝑃 𝜃[𝜃 ∈ 𝐿, 𝑈 ]
1 − 𝛼 is called the confidence coefficient(신뢰도) of the interval
Once the sample is drawn, the realized value of the confidence interval is
𝑙, 𝑢 = (𝐿 𝑥1, 𝑥2, … , 𝑥 𝑛 , 𝑈(𝑥1, 𝑥2, … , 𝑥 𝑛))
2. Confidence Intervals
예제2(구간추정)
베스킨라빈스 파인트의 무게는 정규분포를 따른다고 알고있다.
베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다.
베스킨라빈스 파인트 무게의 평균에 대한 95% 신뢰구간은?
베스킨라빈스 파인트 무게의 모평균이 얼마인지는 모르겠지만,
표본을 보니, 95% 확률로 00~00에 포함되어 있다.
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
2. Confidence Intervals
Confidence Interval for 𝜇 Under Normality
• Suppose 𝑋1, 𝑋2, … , 𝑋 𝑛 are a random sample from a 𝑁 𝜇, 𝜎2 . Let 𝑋 and 𝑆2 denote the sample
mean and sample variance, respectively.
𝑋 is mle of 𝜇 &
𝑛−1
𝑛
𝑆2
is mle of 𝜎2
𝑇 =
𝑋−𝜇
𝑆
𝑛
has a 𝑡-distribution with 𝑛 − 1 degrees of freedom.
1 − 𝛼 = 𝑃(−𝑡 𝛼
2,𝑛−1 < 𝑇 < 𝑡 𝛼
2,𝑛−1)
= 𝑃(−𝑡 𝛼
2,𝑛−1 <
𝑋−𝜇
𝑆
𝑛
< 𝑡 𝛼
2,𝑛−1)
= 𝑃( 𝑋 − 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
< 𝜇 < 𝑋 + 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
)
∴ 𝐿 𝑋1, 𝑋2 , … , 𝑋 𝑛 = 𝑋 − 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
and U 𝑋1, 𝑋2 , … , 𝑋 𝑛 = 𝑋 + 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
2. Confidence Intervals
𝑡-interval and standard error
• Once the sample is drawn, let 𝑥 and 𝑠 denote the realized
values of the statistics 𝑋 and 𝑆. Then 1 − 𝛼 100% conficence
interval for 𝜇 is given by
( 𝑥 − 𝑡 𝛼
2,𝑛−1
𝑠
𝑛
, 𝑥 + 𝑡 𝛼
2,𝑛−1
𝑠
𝑛
)
t-interval
Standard error
2. Confidence Intervals
예제2 풀이
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
𝑥 = 321, 𝑠 = 11.79 𝑛 = 10, 𝑡 𝛼
2,𝑛−1 = 2.262
95% 신뢰구간 = (312.5665, 329.4335)
2. Confidence Intervals
Central Limit Theorem(CLT)
• The last example depends on the normality of the sampled items
• However, this is approximately true even if the sampled items are
not drawn from a normal distribution.
• Theorem 1 (Central Limit Theorem)
• Let 𝑋1, 𝑋2, … , 𝑋 𝑛 is a random sample with mean 𝜇 and finite variance 𝜎2
.
Define a random variable 𝑍 𝑛 =
𝑋−𝜇
𝜎 𝑛
. Then 𝑍 𝑛 →
𝐷
𝑍, where 𝑍~𝑁(0,1).
• 𝑋를 사용하는 경우
• 𝑛 이 충분히 큰 경우, 𝑡-분포를 사용할까? 정규분포를 사용할까?
-> 𝑡-분포 : more conservative
2. Confidence Intervals
Confidence Intervals for Difference in Means
• We compare the means of 𝑋 and 𝑌, denoted by 𝜇1 and 𝜇2,
respectively. Assume that the variances of 𝑋 and Y are finite
and denote them as 𝜎1
2
and 𝜎2
2
, respectively.
• Goal : Estimate ∆= 𝜇1 − 𝜇2.
• What is confidence intervals of ∆= 𝜇1 − 𝜇2?
• Sampling : 𝑋1, 𝑋2, … , 𝑋 𝑛1
& 𝑌1, 𝑌2, … , 𝑌𝑛2
• Let ∆ = 𝑋 − 𝑌. The statistic ∆ is unbiased estimator of ∆.
• 𝑉𝑎𝑟 ∆ =
𝜎1
2
𝑛1
+
𝜎2
2
𝑛2
. By CLT, 𝑍 =
∆−∆
𝑆1
2
𝑛1
+
𝑆2
2
𝑛2
~𝑁(0,12)
( 𝑥 − 𝑦 − 𝑧 𝛼 2
𝑠1
2
𝑛1
+
𝑠2
2
𝑛2
, 𝑥 − 𝑦 + 𝑧 𝛼 2
𝑠1
2
𝑛1
+
𝑠2
2
𝑛2
)
2. Confidence Intervals
Confidence Intervals for Difference in Means
• 𝑋 and 𝑌 are normal with the same variance; i.e., 𝜎2
= 𝜎1
2
= 𝜎2
2
.
• Estimator 𝑆 𝑝
2 of 𝜎2 is weighted average of 𝑆1
2
and 𝑆2
2
.
→ pooled estimator
→ 𝑆 𝑝
2 =
𝑛1−1 𝑆1
2+ 𝑛2−1 𝑆2
2
𝑛1+𝑛2−2
• 𝑇 =
𝑋− 𝑌 − 𝜇1−𝜇2 𝜎 𝑛1
−1+𝑛2
−1
𝑛−2 𝑆 𝑝
2 𝑛−2 𝜎2
has a 𝑡-distribution with 𝑛 − 2 degrees of freedom.
• 1 − 𝛼 100% confidence interval for ∆= 𝜇1 − 𝜇2:
• (( 𝑥- 𝑦) − 𝑡 𝛼 2,𝑛−2 𝑠 𝑝
1
𝑛1
+
1
𝑛2
, 𝑥− 𝑦 + 𝑡 𝛼 2,𝑛−2 𝑠 𝑝
1
𝑛1
+
1
𝑛2
))
2. Confidence Intervals
Order Statistics(순서통계량)
• 전구가 100개 있고, 𝑖 번째 전구의 수명을 확률변수 𝑋𝑖라고 하자.
첫 번째 꺼진 전구의 수명을 𝑌1, 두 번째 꺼진 전구의 수명을 𝑌2,
… , 100번째 꺼진 전구의 수명을 𝑌100 이라고 하자.
• 이렇게 순서대로 얻어진 확률변수를 순서통계량이라고 한다.
• 𝑋1, 𝑋2, … , 𝑋 𝑛이 독립이어도, 𝑌1, 𝑌2, … , 𝑌𝑛은 독립이 아닐 수 있다.
3. Order Statistics
Order Statistics(순서통계량)
• Theorem 2
• 𝑋 : continuous random variable whose pdf 𝑓(𝑥) has support 𝒮 =
𝑎, 𝑏 , where −∞ ≤ 𝑎 < 𝑏 ≤ ∞. Let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 denote the 𝑛
order statistics 𝑋1, 𝑋2, … , 𝑋 𝑛 from 𝑋. Then joint pdf of 𝑌1, 𝑌2, … , 𝑌𝑛 is
given by
𝑔 𝑦1, 𝑦2, … , 𝑦𝑛 =
𝑛! 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓 𝑦𝑛 𝑎 < 𝑦1 < 𝑦2 < ⋯ < 𝑦𝑛 < 𝑏
0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
3. Order Statistics
• Proof
• 𝑔 𝑦1, 𝑦2, … , 𝑦𝑛 = 𝑖=1
𝑛!
𝒥𝑖 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓(𝑦𝑛)
=
𝑛! 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓 𝑦𝑛 𝑎 < 𝑦1 < 𝑦2 < ⋯ < 𝑦𝑛 < 𝑏
0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
∵ 𝒥𝑖 = 1
• Important statistics using order statistics
• Range = 𝑌𝑛 − 𝑌1
• Midrange =
Y1+𝑌𝑛
2
• Median = 𝑌 𝑛+1
2
, 𝑛 𝑖𝑠 𝑜𝑑𝑑.
3. Order Statistics
Quantiles
• Let 𝑋 be a random variable with a continuous cdf 𝐹(𝑥). For
0 < 𝑝 < 1, define the 𝒑th quantile of 𝑋 to be 𝜉 𝑝 = 𝐹−1
(𝑝).
• e.g) 𝜉0.5 is median of 𝑋.
• What is estimator of 𝜉 𝑝?
• Confidence Interval for 𝜉 𝑝?
3. Order Statistics
• Let 𝑋1, 𝑋2, … , 𝑋 𝑛 be a random sample from the distribution of 𝑋
and let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 be the corresponding order statistics. Let
𝑘 be the greatest integer less than or equal to 𝑝(𝑛 + 1).
• 𝐸 𝐹 𝑌𝑘 = 𝑎
𝑏
𝐹 𝑦 𝑘 𝑔 𝑘 𝑦 𝑘 𝑑𝑦 𝑘
= 0
1 𝑛!
𝑘−1 ! 𝑛−𝑘 !
𝑧 𝑘
1 − 𝑧 𝑛−𝑘
𝑑𝑧
=
𝑘
𝑛+1
≈ 𝑝
• We take 𝑌𝑘 as an estimator of the quantile 𝜉 𝑝.
• 𝑌𝑘 is called the 𝒑th sample quantile(or, 100𝒑th percentile of the sample)
• Sample quantiles are useful descriptive statistics.
3. Order Statistics
Application of Quantile : Five-number summary
• A five-number summary of the data consist of the following
sample quantiles
• The minimum(𝑌1), the first quartile(𝑌.25(𝑛+1) = 𝜉.25), the
median(𝑌.50(𝑛+1) = 𝜉.50), the third quartile(𝑌.75(𝑛+1) = 𝜉.75), and the
maximum(𝑌𝑛)
• Use the notation 𝑄1, 𝑄2 , and 𝑄3 to denote, respectively, the first
quartile, median, and third quartile of the sample.
3. Order Statistics
예제2
• 확률변수 𝑋의 샘플 15개 데이터가 다음과 같다.
• The minimum : 𝑦1 = 56
• 𝑄1 = 𝑦4 = 94
• 𝑄2 = 𝑦8 = 102
• 𝑄3 = 𝑦12 = 108
• The maximum : 𝑦15 = 116
56 70 89 94 96 101 102 102
102 105 106 108 110 113 116
3. Order Statistics
Application of Quantile : Box-Whisker plot
• The five-number summary is the basis for a useful and quick plot of the
data.
• ℎ = 1.5(𝑄3 − 𝑄1)
• 𝐥𝐨𝐰𝐞𝐫 𝐟𝐞𝐧𝐜𝐞 𝐿𝐹 = 𝑄1 − ℎ, 𝐮𝐩𝐩𝐞𝐫 𝐟𝐞𝐧𝐜𝐞(𝑈𝐹) = 𝑄3 + ℎ
• Points that lie outside the fences are called potential outliers and they
are denoted by ‘o’ on the boxplot.
• The whiskers protrude from the sides of the box to what are called
adjacent points, which are the points within the fences but closest to
the fences.
3. Order Statistics
56 70 89 94 96 101 102 102
102 105 106 108 110 113 116
The minimum : 𝑦1 = 56
𝑄1 = 𝑦4 = 94
𝑄2 = 𝑦8 = 102
𝑄3 = 𝑦12 = 108
The maximum : 𝑦15 = 116
3. Order Statistics
Application of Quantile : Diagnostic plot
• In practice, we often assume that the data follow a certain
distribution. Such an assumption needs to be checked and
there are many statistical tests which do so
Goodness of fit(위키피디아)
• We discuss one such diagnostic plot in this regard.
• q-q plot
3. Order Statistics
q-q plot
• Suppose 𝑋 is a random variable with cdf 𝐹 𝑥 − 𝑎 𝑏 , where 𝐹(𝑥)
is known but 𝑎 and 𝑏 > 0 may not be.
• Let 𝑍 = (𝑋 − 𝑎) 𝑏. Then, 𝑍 has cdf 𝐹(𝑧). Let 0 < 𝑝 < 1 and let 𝜉 𝑋,𝑝
be the 𝑝th quantile of 𝑋 and 𝜉 𝑍,𝑝 be the 𝑝th quantile of 𝑍.
• 𝑝 = 𝑃 𝑋 ≤ 𝜉 𝑋,𝑝 = 𝑃[𝑍 ≤
𝜉 𝑋,𝑝−𝑎
𝑏
]
⇒ 𝜉 𝑋,𝑝 = 𝑏𝜉 𝑍,𝑝 + 𝑎
• Thus, the quantiles of 𝑋 are linearly related to the quantiles of 𝑍.
3. Order Statistics
q-q plot
• 𝑋1, 𝑋2, … , 𝑋 𝑛 : Random sample from 𝑋 ⇒ 𝑌1 < ⋯ < 𝑌𝑛 the order statistics.
• For 𝑘 = 1, … , 𝑛, let 𝑝 𝑘 =
𝑘
𝑛+1
. ⇒ 𝑌𝑘 is an estimator of 𝜉 𝑋,𝑝.
• Denote the corresponding quantiles of the cdf 𝐹 𝑧 by 𝜉 𝑍,𝑝 = 𝐹−1(𝑝 𝑘)
• The plot of 𝑌𝑘 versus 𝜉 𝑍,𝑝 is called a q-q plot.
3. Order Statistics
• Figure 4.1 contain q-q plots of three different
Distributions
• Panel C appears to be most linear. We may assume that the data were generated from a
Laplace distribution
3. Order Statistics
Confidence Intervals for Quantiles
• For a sample size 𝑛 on 𝑋, let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 be the order statistics. Let 𝑘 =
[ 𝑛 + 1 𝑝]. Then the 100𝑝th sample percentile 𝑌𝑘 is a point estimator of 𝜉 𝑝.
• We now derive distribution free confidence interval for 𝜉 𝑝.
• Let 𝑖 < 𝑛 + 1 𝑝 < 𝑗, and consider the order statistics 𝑌𝑖 < 𝑌𝑗 and the event
𝑌𝑖 < 𝜉 𝑝 ∩ 𝑌𝑗 > 𝜉 𝑝 = [적어도 i개는 𝜉 𝑝보다 작고 𝑗번째는 𝜉 𝑝보다 큰 경우]
• The probability of success is 𝑃 𝑋 < 𝜉 𝑝 = 𝐹 𝜉 𝑝 = 𝑝.
𝑃 𝑌𝑖 < 𝜉 𝑝 < 𝑌𝑗 =
𝑤=𝑖
𝑗−1
𝑛
𝑤
𝑝 𝑤
1 − 𝑝 𝑛−𝑤
(= 0.95 → 𝑌𝑖, 𝑌𝑗 )
3. Order Statistics
Example:
Confidence Interval for the Median.
• 𝜉.50 ∶ median of 𝐹(𝑥), 𝑄2 ∶ sample median − point estimator 𝑜𝑓𝜉.50
• Select 0 < 𝛼 < 1 and Take 𝑐 𝛼 2 to be the
𝛼
2
𝑡ℎ quantile of a binomial S ~ 𝑏(𝑛,
1
2
).
• 𝑃 𝑆 ≤ 𝑐 𝛼 2 = 𝑃 𝑆 ≥ 𝑛 − 𝑐 𝛼 2 = 𝛼 2.
• Thus it follows that
𝑃 𝑌𝑐 𝛼 2+1 < 𝜉.50 < 𝑌𝑛−𝑐 𝛼 2
= 1 − 𝛼
• Hence, when the sample is drawn,
(𝑦𝑐 𝛼 2+1, 𝑦𝑛−𝑐 𝛼 2
) is a (1 − 𝛼)100% confidence interval for 𝜉.50.
3. Order Statistics
예제3
• 모기 4마리를 채집해서 모기 4마리의 수명을 관찰했다. 모기의
수명이 평균이 1인 지수분포를 따를 때, 제일 오래 산 모기의 수
명이 3 이상일 확률은?
• Let 𝑌1 < 𝑌2 < 𝑌3 < 𝑌4 be the order statistics of a random
sample of size 4 from the distribution having pdf 𝑓 𝑥 = 𝑒−𝑥,
0 < 𝑥 < ∞. Find 𝑃 3 ≤ 𝑌4 .
• 𝑃 𝑌4 ≤ 𝑡 = 1 − 𝑒−𝑡 4
∴ 𝑃 𝑌4 ≥ 3 = 1 − 1 − 𝑒−3 4 = 0.1848
3. Order Statistics
예제4
• 모기 4마리를 채집해서 모기 4마리의 수명을 관찰했다. 모기의
수명이 0에서 1의 값을 가지는 균등분포를 따를 때, 제일 오래
산 모기와 제일 빨리 죽은 모기의 수명의 차이가 0.5 미만일 확
률은?
• Let 𝑌1 < 𝑌2 < 𝑌3 < 𝑌4 be the order statistics of a random
sample of size 4 from the distribution having pdf 𝑓 𝑥 = 1,
0 < 𝑥 < 1. Find the probability that the range of the random
sample is less then
1
2
.
3. Order Statistics
• 𝑠𝑜𝑙) 𝑃 𝑌4 − 𝑌1 >
1
2
= 0
1
2
𝑦1+
1
2
1
𝑔1,4(𝑦1, 𝑦4)𝑑𝑦4 𝑑𝑦1
• 𝑔1,4 𝑦1, 𝑦4 = 𝑦1
𝑦4
𝑦2
𝑦4
𝑔 𝑦1, 𝑦2, 𝑦3, 𝑦4 𝑑𝑦3 𝑑𝑦2 = 12 𝑦4 − 𝑦1
2
• 𝑃 𝑌4 − 𝑌1 ≤
1
2
= 1 − 0
1
2
𝑦1+
1
2
1
12 𝑦4 − 𝑦1
2 𝑑𝑦4 𝑑𝑦1 =
11
16
3. Order Statistics
Introduction to Hypothesis Testing
4 Introduction to Hypothesis Testing
Statistical Inference
Estimation
Hypothesis Testing
Point estimation
Interval estimation
예제5(점추정/가설검정)
베스킨라빈스 파인트의 무게는 정규분포를 따른다고 가정한다. 베스킨라빈스에 10
번을 가서, 무게를 재봤더니 다음과 같았다.
MLE에 의한 파인트 무게의 모평균의 추정값은??
표본 평균 : 321g -> 모평균의 추정값이 된다!
만약 회사 홈페이지에 310g이라고 써져있다면?
가설검정
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
4 Introduction to Hypothesis Testing
• 귀무가설 : 파인트 무게는 평균 310g이다.
• 연구가설 : 파인트 무게는 평균 310g이 아니다.
• 모평균의 추정값 : 321g -> 귀무가설 기각 (X)
• 기각역 설정 -> 추정값이 기각역에 포함??(O)
1 2 3 4 5 6 7 8 9 10 평균
310 315 330 320 325 325 310 315 320 340 321
4 Introduction to Hypothesis Testing
모분포 : 평균 100, 분산 30
Sampling
표본 평균 : 90
50개
표본 평균 : 80
표본 평균 : 40
표본 평균 : 10
표본 평균 : 1
표본 평균 : 150
귀무가설
귀무가설의 모평균에서 멀어질수록
귀무가설이 틀렸다는 생각이 커진다.
WHY?
50개 뽑았으면 우연이 아니라 필연!
4 Introduction to Hypothesis Testing
Hypothesis
• 𝐻0 : Null hypothesis(귀무가설)
• Represents no change or no difference from the past
• 𝐻1 : Alternative hypothesis(대립가설/연구가설)
• Represents change or difference
• Research worker’s hypothesis
𝐻0 ∶ 𝜃 ∈ 𝜔0 versus 𝐻1 ∶ 𝜃 ∈ 𝜔1
(where 𝜔0 ∪ 𝜔1 = Ω , and 𝜔0 ∩ 𝜔1 = ∅)
• The decision rule to take 𝐻0 or 𝐻0 is based on a sample 𝑋1, 𝑋2, … , 𝑋 𝑛
from 𝑋
• Hence, the decision could wrong.
4 Introduction to Hypothesis Testing
Decision Rule
• 𝑋1, … , 𝑋 𝑛 : random sample from the distribution of 𝑋.
• Consider testing the hypotheses :
• 𝐻0 ∶ 𝜃 ∈ 𝜔0 versus 𝐻1 ∶ 𝜃 ∈ 𝜔1
• Denote the space of the sample by 𝒟 = space{(𝑋1, … , 𝑋 𝑛)}.
• Definition 5 (Statistical Test, Critical Region)
• A test of 𝐻0 versus 𝐻1 is based on a subset 𝐶 of 𝒟.
• This set 𝐶 is called the critical region(기각역) and its corresponding
decision rule is:
Reject 𝐻0 if 𝑋1, … , 𝑋 𝑛 ∈ 𝐶
Retain 𝐻0 if 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 𝑐
4 Introduction to Hypothesis Testing
• the decision could wrong.
• Type 1, Type 2 error
• Goal : Select a critical region from all possible critical regions which minimizes
the probabilities of these two errors.
• in general, not possible (trade-off)
• Procedure
• Select critical regions which bound the probability of Type 1 error
• Among these critical regions we try to select one which minimizes Type 2 error.
4 Introduction to Hypothesis Testing
• Definition 5 (Critical region of size 𝛼)
• We say a critical region 𝐶 is of size 𝛼 if
𝛼 = max
𝜃∈𝜔0
𝑃 𝜃[ 𝑋1, … 𝑋 𝑛 ∈ 𝐶]
• Over all critical region of size 𝛼, we want to consider critical
regions which have lower probabilities of Type 2 error.
• 𝑃 𝜃 Type 2 Error = 𝑃 𝜃 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 𝑐
, 𝜃 ∈ 𝜔1
• we desire to maximize 𝑃 𝜃[ 𝑋1, … , 𝑋 𝑛 ∈ 𝐶]. (power of the test at 𝜃)
• Definition 6 (Power function)
• Power function of a critical region to be
𝛾 𝐶 𝜃 = 𝑃 𝜃 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 ; 𝜃 ∈ 𝜔1
4 Introduction to Hypothesis Testing
Better critical region
• Given two critical regions 𝐶1 and 𝐶2 which are both of size 𝛼,
𝐶1 is better than 𝐶2 if 𝛾 𝐶1
𝜃 ≥ 𝛾 𝐶2
(𝜃) for all 𝜃 ∈ 𝜔1.
• Type1 에러의 최대값이 𝛼로 같으면, Type2 에러가 작은 것이 더 좋은
기각역이다.
4 Introduction to Hypothesis Testing
Test for a Binomial Proportion of Success
• 𝑋~𝐵𝑒𝑟(𝑝), 𝑝0 : probability of dying with some standard treatment
• We want to test at size 𝛼,
𝐻0 ∶ 𝑝 = 𝑝0 versus 𝐻1 ∶ 𝑝 < 𝑝0
• Let 𝑋1, … , 𝑋 𝑛 be a random sample from the distribution of X and
let 𝑆 = 𝑖=1
𝑛
𝑋𝑖 .
• An intuitive decision rule is :
Reject 𝐻0 in favor of 𝐻1 if 𝑆 ≤ 𝑘,
where 𝑘 is such that 𝛼 = 𝑃 𝐻0
[𝑆 ≤ 𝑘].
• For example, suppose 𝑛 = 20, 𝑝0 = 0.7, and 𝛼=0.15
• 𝑃 𝐻0
𝑆 ≤ 11 = 0.1133 and 𝑃 𝐻0
𝑆 ≤ 12 = 0.2277 ⋯ 𝑘 = 11? 12?
• On the conservative side, choose 𝑘 to be 11 and 𝛼 = 0.1133.
• 𝛾 𝑝 = 𝑃𝑝 𝑆 ≤ 𝑘 , 𝑝 < 𝑝0
4 Introduction to Hypothesis Testing
4 Introduction to Hypothesis Testing
Nomenclature
• 𝐻0 ∶ 𝑝 = 𝑝0 completely specifies the underlying distribution, it
is called a simple hypothesis. ( 𝜔0 = 1)
• 𝐻1 ∶ 𝑝 < 𝑝0 is composite hypothesis.
• Frequently, 𝛼 is also called the significance level(유의수준) of
the test
• Or, ”maximum of probabilities of committing an error of Type 1 error”
• Or, “maximum of the power of the test when 𝐻0 is true”
4 Introduction to Hypothesis Testing
𝒕-test : Test for 𝜇 Under Normality
• Let 𝑋 have a 𝑁(𝜇, 𝜎2) distribution. Consider the hypotheses
𝐻0 ∶ 𝜇 = 𝜇0 𝑣𝑒𝑟𝑠𝑢𝑠 𝐻1 ∶ 𝜇 > 𝜇0
• Assume that the desired size of the test is 𝛼, for 0 < 𝛼 < 1.
• Our intuitive rejection rule is to reject 𝐻0 if 𝑋 is much larger than 𝜇0.
• 𝑇 =
𝑋−𝜇0
𝑆 𝑛
, 𝑡-distribution with 𝑛-1 degree of freedom. It follows that this rejection rule
has exact level 𝛼 :
Reject 𝐻0 if 𝑇 =
𝑋 − 𝜇0
𝑆 𝑛
≥ 𝑡 𝛼,𝑛−1, where 𝛼 = 𝑃[𝑇 > 𝑡 𝛼,𝑛−1]
4 Introduction to Hypothesis Testing
• Large sample rule ??
• In practice, we may not be willing to assume that the
population is normal. Usually 𝑡-critical values are larger than
𝑧-critical values
• Hence, the 𝑡-test is conservative relative to the large sample
test. So, in practice, many statisticians often use the 𝑡-test.
4 Introduction to Hypothesis Testing
예제6(가설검정 𝑡-test)
베스킨라빈스 파인트의 무게는 모평균이 310g 인 정규분포를 따
른다고 알고있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니
다음과 같았다.
가설 설정 :
• 귀무가설(𝐻0) : 𝜇 = 310
• 연구가설(𝐻1) : 𝜇 > 310
1 2 3 4 5 6 7 8 9 10 평균
310 315 330 320 325 325 310 315 320 340 321
4 Introduction to Hypothesis Testing
• 기각역 설정(유의수준(𝛼) = 5%)
• Intuitive rejection rule :
• 𝑋 much larger than 310 or much smaller than 310
• 𝑇 =
𝑋−310
𝑆 10
≥ 𝑡0.05,9
• Critical region of size 0.05
• 𝐶 = { 𝑥 ≥ 312.1170 }
• 𝑥 = 321 -> 귀묵가설 기각
• Conclusion
• 베스킨라빈스 파인트 모평균은 310g이 아니고, 310g 보다 크다.
4 Introduction to Hypothesis Testing
Randomized tests and 𝑝-value
• Let 𝑋1, … , 𝑋10 be random sample of size 𝑛=10 from a Poisson
distribution with mean 𝜃.
• A critical region for testing 𝐻0 ∶ 𝜃 = 0.1 against 𝐻1 ∶ 𝜃 > 0.1 is
given by 𝑌 = 𝑖=1
10
𝑋𝑖 ≥ 3. The statistic 𝑌 has a Poisson distribution
with mean 10𝜃.
• Significance level of the test is
𝑃 𝑌 ≥ 3 = 1 − 𝑃 𝑌 ≤ 2 = 1 − 0.920 = 0.080
• If the critical region defined by 𝑖=1
10
𝑋𝑖 ≥ 4 is used, the significance level is
𝛼 = 𝑃 𝑌 ≥ 4 = 1 − 𝑃 𝑌 ≤ 3 = 0.019
How we achieve a significance level of 𝛼 = 0.05??
4 Introduction to Hypothesis Testing
• Let 𝑊 have a Bernoulli distribution with probability of success
equal to
𝑃 𝑊 = 1 =
0.050 − 0.019
0.080 − 0.019
=
31
61
• Assume that 𝑊 is selected independently of the sample. Consider
the rejection rule
Reject 𝐻0 if 1
10
𝑥𝑖 ≥ 4 or if 1
10
𝑥𝑖 = 3 and 𝑊 = 1.
• The significance level of this rule is
𝑃 𝐻0
𝑌 ≥ 4 + 𝑃 𝑌 = 3 ∩ 𝑊 = 1
= 𝑃 𝐻0
𝑌 ≥ 4 + 𝑃 𝐻0
𝑌 = 3 𝑃 𝑊 = 1
= 0.019 + 0.061
31
61
= 0.05
• The process of performing the auxiliary experiment is referred to
as a randomized test.
4 Introduction to Hypothesis Testing
Observed Significance Level
• Not many statisticians like randomized tests in practice,
because the use of them means that two statisticians could
make the same assumptions, observe the same data, and yet
make different decisions.
• As a matter of fact, many statisticians report what are
commonly called observed significance level or 𝒑-values.
유의수준을 정했는데 그에 따른 기각역을 정할 수 없으면 그냥 p-value를 살피자.
4 Introduction to Hypothesis Testing
모분포 : 평균 100, 분산 30
Sampling
50개
표본 평균 : 30
표본 평균 : 180
귀무가설
두 표본이 얼마나 귀무가설을
벗어나는지 측정하고 싶다.
표본평균 30이 뽑히는 것만큼
극단적으로 관찰될 확률은?
표본평균 180이 뽑히는 것만큼
극단적으로 관찰될 확률은?
유의수준을 정해준 것은 귀무가설을 얼마나 벗어나면 기각을 시킬지 미리 정해준것!
4 Introduction to Hypothesis Testing
P-value
• 𝑝-value is the observed “tail” probability of a statistic being at
least as extreme as the particular observed value when 𝐻0 is
true.
• If 𝑌 = 𝑢(𝑋1, 𝑋2, … , 𝑋 𝑛) is the statistic to be used in a test of
𝐻0 and if the critical region is of the form
𝑢 𝑥1, 𝑥2, … , 𝑥 𝑛 ≤ 𝑐,
an observed value 𝑢 𝑥1, 𝑥2, … , 𝑥 𝑛 = 𝑑 would mean that the
𝑝-value = 𝑃(𝑌 ≤ 𝑑; 𝐻0)
4 Introduction to Hypothesis Testing
예제7
• 𝑋1, 𝑋2, … 𝑋25 a random sample from 𝑁(𝜇, 𝜎2 = 4). 𝑋~𝑁(𝜇, 0.16)
• Test 𝐻0 ∶ 𝜇 = 77 against the one-sided alternative hypothesis
𝐻1 ∶ 𝜇 < 77
• 𝑥 = 76.1
• 𝑧 − 𝑠𝑐𝑜𝑟𝑒 : -2.25
• 𝑝 − 𝑣𝑎𝑙𝑢𝑒 ∶ ∅ −2.25 = 0.012
• If we were using a significance level of 𝛼 = 0.05, we would
reject 𝐻0 and accept 𝐻1 ∶ 𝜇 < 77 because 0.012 < 0.05
4 Introduction to Hypothesis Testing

More Related Content

What's hot

Estimating a Population Standard Deviation or Variance
Estimating a Population Standard Deviation or Variance Estimating a Population Standard Deviation or Variance
Estimating a Population Standard Deviation or Variance
Long Beach City College
 
02a one sample_t-test
02a one sample_t-test02a one sample_t-test
02a one sample_t-test
Madhusudhanan Balakumar
 
Statistical Estimation
Statistical Estimation Statistical Estimation
Statistical Estimation
Remyagharishs
 
Statistical parameters
Statistical parametersStatistical parameters
Statistical parameters
Burdwan University
 
Hypothesis and Test
Hypothesis and TestHypothesis and Test
Hypothesis and Test
Avjinder (Avi) Kaler
 
Linear Regression
Linear Regression Linear Regression
Linear Regression
Rupak Roy
 
statistical estimation
statistical estimationstatistical estimation
statistical estimation
Amish Akbar
 
Population and sample mean
Population and sample meanPopulation and sample mean
Population and sample mean
Avjinder (Avi) Kaler
 
Hypo
HypoHypo
Estimation
EstimationEstimation
Estimation
rishi.indian
 
C2 st lecture 11 the t-test handout
C2 st lecture 11   the t-test handoutC2 st lecture 11   the t-test handout
C2 st lecture 11 the t-test handout
fatima d
 
Statistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionStatistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling Distribution
Dexlab Analytics
 
Basic statistics 1
Basic statistics  1Basic statistics  1
Basic statistics 1
Kumar P
 
Normal and standard normal distribution
Normal and standard normal distributionNormal and standard normal distribution
Normal and standard normal distribution
Avjinder (Avi) Kaler
 
Descriptive statistics and graphs
Descriptive statistics and graphsDescriptive statistics and graphs
Descriptive statistics and graphs
Avjinder (Avi) Kaler
 
law of large number and central limit theorem
 law of large number and central limit theorem law of large number and central limit theorem
law of large number and central limit theorem
lovemucheca
 
A Lecture on Sample Size and Statistical Inference for Health Researchers
A Lecture on Sample Size and Statistical Inference for Health ResearchersA Lecture on Sample Size and Statistical Inference for Health Researchers
A Lecture on Sample Size and Statistical Inference for Health Researchers
Dr Arindam Basu
 
Sampling Distributions and Estimators
Sampling Distributions and EstimatorsSampling Distributions and Estimators
Sampling Distributions and Estimators
Long Beach City College
 
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik   statistik deskriptif 2 numerikEe184405 statistika dan stokastik   statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
yusufbf
 
Point Estimation
Point Estimation Point Estimation

What's hot (20)

Estimating a Population Standard Deviation or Variance
Estimating a Population Standard Deviation or Variance Estimating a Population Standard Deviation or Variance
Estimating a Population Standard Deviation or Variance
 
02a one sample_t-test
02a one sample_t-test02a one sample_t-test
02a one sample_t-test
 
Statistical Estimation
Statistical Estimation Statistical Estimation
Statistical Estimation
 
Statistical parameters
Statistical parametersStatistical parameters
Statistical parameters
 
Hypothesis and Test
Hypothesis and TestHypothesis and Test
Hypothesis and Test
 
Linear Regression
Linear Regression Linear Regression
Linear Regression
 
statistical estimation
statistical estimationstatistical estimation
statistical estimation
 
Population and sample mean
Population and sample meanPopulation and sample mean
Population and sample mean
 
Hypo
HypoHypo
Hypo
 
Estimation
EstimationEstimation
Estimation
 
C2 st lecture 11 the t-test handout
C2 st lecture 11   the t-test handoutC2 st lecture 11   the t-test handout
C2 st lecture 11 the t-test handout
 
Statistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionStatistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling Distribution
 
Basic statistics 1
Basic statistics  1Basic statistics  1
Basic statistics 1
 
Normal and standard normal distribution
Normal and standard normal distributionNormal and standard normal distribution
Normal and standard normal distribution
 
Descriptive statistics and graphs
Descriptive statistics and graphsDescriptive statistics and graphs
Descriptive statistics and graphs
 
law of large number and central limit theorem
 law of large number and central limit theorem law of large number and central limit theorem
law of large number and central limit theorem
 
A Lecture on Sample Size and Statistical Inference for Health Researchers
A Lecture on Sample Size and Statistical Inference for Health ResearchersA Lecture on Sample Size and Statistical Inference for Health Researchers
A Lecture on Sample Size and Statistical Inference for Health Researchers
 
Sampling Distributions and Estimators
Sampling Distributions and EstimatorsSampling Distributions and Estimators
Sampling Distributions and Estimators
 
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik   statistik deskriptif 2 numerikEe184405 statistika dan stokastik   statistik deskriptif 2 numerik
Ee184405 statistika dan stokastik statistik deskriptif 2 numerik
 
Point Estimation
Point Estimation Point Estimation
Point Estimation
 

Similar to Elementary statistical inference1

Inferential Statistics.pdf
Inferential Statistics.pdfInferential Statistics.pdf
Inferential Statistics.pdf
Shivakumar B N
 
Resampling methods
Resampling methodsResampling methods
Resampling methods
Setia Pramana
 
1.1 course notes inferential statistics
1.1 course notes inferential statistics1.1 course notes inferential statistics
1.1 course notes inferential statistics
Djamel Bob
 
Ch3_Statistical Analysis and Random Error Estimation.pdf
Ch3_Statistical Analysis and Random Error Estimation.pdfCh3_Statistical Analysis and Random Error Estimation.pdf
Ch3_Statistical Analysis and Random Error Estimation.pdf
Vamshi962726
 
Identifying the sampling distribution module5
Identifying the sampling distribution module5Identifying the sampling distribution module5
Identifying the sampling distribution module5
REYEMMANUELILUMBA
 
U unit8 ksb
U unit8 ksbU unit8 ksb
U unit8 ksb
Akhilesh Deshpande
 
Chapter one on sampling distributions.ppt
Chapter one on sampling distributions.pptChapter one on sampling distributions.ppt
Chapter one on sampling distributions.ppt
FekaduAman
 
Inorganic CHEMISTRY
Inorganic CHEMISTRYInorganic CHEMISTRY
Inorganic CHEMISTRY
Saikumar raja
 
Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...
Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...
Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...
Daniel Katz
 
STATISTIC ESTIMATION
STATISTIC ESTIMATIONSTATISTIC ESTIMATION
STATISTIC ESTIMATION
Smruti Ranjan Parida
 
Lect 3 background mathematics for Data Mining
Lect 3 background mathematics for Data MiningLect 3 background mathematics for Data Mining
Lect 3 background mathematics for Data Mining
hktripathy
 
Lect 3 background mathematics
Lect 3 background mathematicsLect 3 background mathematics
Lect 3 background mathematics
hktripathy
 
03 Data Mining Techniques
03 Data Mining Techniques03 Data Mining Techniques
03 Data Mining Techniques
Valerii Klymchuk
 
Data Ananlysis lecture 7 Simon Fraser University
Data Ananlysis lecture 7 Simon Fraser UniversityData Ananlysis lecture 7 Simon Fraser University
Data Ananlysis lecture 7 Simon Fraser University
soniyamarghani
 
Statistik dan Probabilitas Yuni Yamasari 2.pptx
Statistik dan Probabilitas Yuni Yamasari 2.pptxStatistik dan Probabilitas Yuni Yamasari 2.pptx
Statistik dan Probabilitas Yuni Yamasari 2.pptx
AisyahLailia
 
Unit3
Unit3Unit3
Machine learning mathematicals.pdf
Machine learning mathematicals.pdfMachine learning mathematicals.pdf
Machine learning mathematicals.pdf
King Khalid University
 
UNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptxUNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptx
kalpana413121
 
Deep Learning Theory Seminar (Chap 3, part 2)
Deep Learning Theory Seminar (Chap 3, part 2)Deep Learning Theory Seminar (Chap 3, part 2)
Deep Learning Theory Seminar (Chap 3, part 2)
Sangwoo Mo
 
Bayesian Neural Networks
Bayesian Neural NetworksBayesian Neural Networks
Bayesian Neural Networks
Natan Katz
 

Similar to Elementary statistical inference1 (20)

Inferential Statistics.pdf
Inferential Statistics.pdfInferential Statistics.pdf
Inferential Statistics.pdf
 
Resampling methods
Resampling methodsResampling methods
Resampling methods
 
1.1 course notes inferential statistics
1.1 course notes inferential statistics1.1 course notes inferential statistics
1.1 course notes inferential statistics
 
Ch3_Statistical Analysis and Random Error Estimation.pdf
Ch3_Statistical Analysis and Random Error Estimation.pdfCh3_Statistical Analysis and Random Error Estimation.pdf
Ch3_Statistical Analysis and Random Error Estimation.pdf
 
Identifying the sampling distribution module5
Identifying the sampling distribution module5Identifying the sampling distribution module5
Identifying the sampling distribution module5
 
U unit8 ksb
U unit8 ksbU unit8 ksb
U unit8 ksb
 
Chapter one on sampling distributions.ppt
Chapter one on sampling distributions.pptChapter one on sampling distributions.ppt
Chapter one on sampling distributions.ppt
 
Inorganic CHEMISTRY
Inorganic CHEMISTRYInorganic CHEMISTRY
Inorganic CHEMISTRY
 
Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...
Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...
Quantitative Methods for Lawyers - Class #10 - Binomial Distributions, Normal...
 
STATISTIC ESTIMATION
STATISTIC ESTIMATIONSTATISTIC ESTIMATION
STATISTIC ESTIMATION
 
Lect 3 background mathematics for Data Mining
Lect 3 background mathematics for Data MiningLect 3 background mathematics for Data Mining
Lect 3 background mathematics for Data Mining
 
Lect 3 background mathematics
Lect 3 background mathematicsLect 3 background mathematics
Lect 3 background mathematics
 
03 Data Mining Techniques
03 Data Mining Techniques03 Data Mining Techniques
03 Data Mining Techniques
 
Data Ananlysis lecture 7 Simon Fraser University
Data Ananlysis lecture 7 Simon Fraser UniversityData Ananlysis lecture 7 Simon Fraser University
Data Ananlysis lecture 7 Simon Fraser University
 
Statistik dan Probabilitas Yuni Yamasari 2.pptx
Statistik dan Probabilitas Yuni Yamasari 2.pptxStatistik dan Probabilitas Yuni Yamasari 2.pptx
Statistik dan Probabilitas Yuni Yamasari 2.pptx
 
Unit3
Unit3Unit3
Unit3
 
Machine learning mathematicals.pdf
Machine learning mathematicals.pdfMachine learning mathematicals.pdf
Machine learning mathematicals.pdf
 
UNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptxUNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptx
 
Deep Learning Theory Seminar (Chap 3, part 2)
Deep Learning Theory Seminar (Chap 3, part 2)Deep Learning Theory Seminar (Chap 3, part 2)
Deep Learning Theory Seminar (Chap 3, part 2)
 
Bayesian Neural Networks
Bayesian Neural NetworksBayesian Neural Networks
Bayesian Neural Networks
 

More from SEMINARGROOT

Metric based meta_learning
Metric based meta_learningMetric based meta_learning
Metric based meta_learning
SEMINARGROOT
 
Sampling method : MCMC
Sampling method : MCMCSampling method : MCMC
Sampling method : MCMC
SEMINARGROOT
 
Demystifying Neural Style Transfer
Demystifying Neural Style TransferDemystifying Neural Style Transfer
Demystifying Neural Style Transfer
SEMINARGROOT
 
Towards Deep Learning Models Resistant to Adversarial Attacks.
Towards Deep Learning Models Resistant to Adversarial Attacks.Towards Deep Learning Models Resistant to Adversarial Attacks.
Towards Deep Learning Models Resistant to Adversarial Attacks.
SEMINARGROOT
 
The ways of node embedding
The ways of node embeddingThe ways of node embedding
The ways of node embedding
SEMINARGROOT
 
Graph Convolutional Network
Graph  Convolutional NetworkGraph  Convolutional Network
Graph Convolutional Network
SEMINARGROOT
 
Denoising With Frequency Domain
Denoising With Frequency DomainDenoising With Frequency Domain
Denoising With Frequency Domain
SEMINARGROOT
 
Bayesian Statistics
Bayesian StatisticsBayesian Statistics
Bayesian Statistics
SEMINARGROOT
 
Coding Test Review 3
Coding Test Review 3Coding Test Review 3
Coding Test Review 3
SEMINARGROOT
 
Time Series Analysis - ARMA
Time Series Analysis - ARMATime Series Analysis - ARMA
Time Series Analysis - ARMA
SEMINARGROOT
 
Differential Geometry for Machine Learning
Differential Geometry for Machine LearningDifferential Geometry for Machine Learning
Differential Geometry for Machine Learning
SEMINARGROOT
 
Generative models : VAE and GAN
Generative models : VAE and GANGenerative models : VAE and GAN
Generative models : VAE and GAN
SEMINARGROOT
 
Effective Python
Effective PythonEffective Python
Effective Python
SEMINARGROOT
 
Understanding Blackbox Prediction via Influence Functions
Understanding Blackbox Prediction via Influence FunctionsUnderstanding Blackbox Prediction via Influence Functions
Understanding Blackbox Prediction via Influence Functions
SEMINARGROOT
 
Attention Is All You Need
Attention Is All You NeedAttention Is All You Need
Attention Is All You Need
SEMINARGROOT
 
Attention
AttentionAttention
Attention
SEMINARGROOT
 
WWW 2020 XAI Tutorial Review
WWW 2020 XAI Tutorial ReviewWWW 2020 XAI Tutorial Review
WWW 2020 XAI Tutorial Review
SEMINARGROOT
 
Coding test review 2
Coding test review 2Coding test review 2
Coding test review 2
SEMINARGROOT
 
Locality sensitive hashing
Locality sensitive hashingLocality sensitive hashing
Locality sensitive hashing
SEMINARGROOT
 
Coding Test Review1
Coding Test Review1Coding Test Review1
Coding Test Review1
SEMINARGROOT
 

More from SEMINARGROOT (20)

Metric based meta_learning
Metric based meta_learningMetric based meta_learning
Metric based meta_learning
 
Sampling method : MCMC
Sampling method : MCMCSampling method : MCMC
Sampling method : MCMC
 
Demystifying Neural Style Transfer
Demystifying Neural Style TransferDemystifying Neural Style Transfer
Demystifying Neural Style Transfer
 
Towards Deep Learning Models Resistant to Adversarial Attacks.
Towards Deep Learning Models Resistant to Adversarial Attacks.Towards Deep Learning Models Resistant to Adversarial Attacks.
Towards Deep Learning Models Resistant to Adversarial Attacks.
 
The ways of node embedding
The ways of node embeddingThe ways of node embedding
The ways of node embedding
 
Graph Convolutional Network
Graph  Convolutional NetworkGraph  Convolutional Network
Graph Convolutional Network
 
Denoising With Frequency Domain
Denoising With Frequency DomainDenoising With Frequency Domain
Denoising With Frequency Domain
 
Bayesian Statistics
Bayesian StatisticsBayesian Statistics
Bayesian Statistics
 
Coding Test Review 3
Coding Test Review 3Coding Test Review 3
Coding Test Review 3
 
Time Series Analysis - ARMA
Time Series Analysis - ARMATime Series Analysis - ARMA
Time Series Analysis - ARMA
 
Differential Geometry for Machine Learning
Differential Geometry for Machine LearningDifferential Geometry for Machine Learning
Differential Geometry for Machine Learning
 
Generative models : VAE and GAN
Generative models : VAE and GANGenerative models : VAE and GAN
Generative models : VAE and GAN
 
Effective Python
Effective PythonEffective Python
Effective Python
 
Understanding Blackbox Prediction via Influence Functions
Understanding Blackbox Prediction via Influence FunctionsUnderstanding Blackbox Prediction via Influence Functions
Understanding Blackbox Prediction via Influence Functions
 
Attention Is All You Need
Attention Is All You NeedAttention Is All You Need
Attention Is All You Need
 
Attention
AttentionAttention
Attention
 
WWW 2020 XAI Tutorial Review
WWW 2020 XAI Tutorial ReviewWWW 2020 XAI Tutorial Review
WWW 2020 XAI Tutorial Review
 
Coding test review 2
Coding test review 2Coding test review 2
Coding test review 2
 
Locality sensitive hashing
Locality sensitive hashingLocality sensitive hashing
Locality sensitive hashing
 
Coding Test Review1
Coding Test Review1Coding Test Review1
Coding Test Review1
 

Recently uploaded

4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...
Social Samosa
 
A presentation that explain the Power BI Licensing
A presentation that explain the Power BI LicensingA presentation that explain the Power BI Licensing
A presentation that explain the Power BI Licensing
AlessioFois2
 
Intelligence supported media monitoring in veterinary medicine
Intelligence supported media monitoring in veterinary medicineIntelligence supported media monitoring in veterinary medicine
Intelligence supported media monitoring in veterinary medicine
AndrzejJarynowski
 
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
Walaa Eldin Moustafa
 
Build applications with generative AI on Google Cloud
Build applications with generative AI on Google CloudBuild applications with generative AI on Google Cloud
Build applications with generative AI on Google Cloud
Márton Kodok
 
Experts live - Improving user adoption with AI
Experts live - Improving user adoption with AIExperts live - Improving user adoption with AI
Experts live - Improving user adoption with AI
jitskeb
 
DSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelinesDSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelines
Timothy Spann
 
DATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docx
DATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docxDATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docx
DATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docx
SaffaIbrahim1
 
原版一比一多伦多大学毕业证(UofT毕业证书)如何办理
原版一比一多伦多大学毕业证(UofT毕业证书)如何办理原版一比一多伦多大学毕业证(UofT毕业证书)如何办理
原版一比一多伦多大学毕业证(UofT毕业证书)如何办理
mkkikqvo
 
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
Timothy Spann
 
一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理
一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理
一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理
nuttdpt
 
一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理
bmucuha
 
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Aggregage
 
Monthly Management report for the Month of May 2024
Monthly Management report for the Month of May 2024Monthly Management report for the Month of May 2024
Monthly Management report for the Month of May 2024
facilitymanager11
 
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
v7oacc3l
 
一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理
一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理
一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理
hyfjgavov
 
Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...
Bill641377
 
UofT毕业证如何办理
UofT毕业证如何办理UofT毕业证如何办理
UofT毕业证如何办理
exukyp
 
一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理
一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理
一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理
xclpvhuk
 
一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理
一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理
一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理
bopyb
 

Recently uploaded (20)

4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...
 
A presentation that explain the Power BI Licensing
A presentation that explain the Power BI LicensingA presentation that explain the Power BI Licensing
A presentation that explain the Power BI Licensing
 
Intelligence supported media monitoring in veterinary medicine
Intelligence supported media monitoring in veterinary medicineIntelligence supported media monitoring in veterinary medicine
Intelligence supported media monitoring in veterinary medicine
 
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
 
Build applications with generative AI on Google Cloud
Build applications with generative AI on Google CloudBuild applications with generative AI on Google Cloud
Build applications with generative AI on Google Cloud
 
Experts live - Improving user adoption with AI
Experts live - Improving user adoption with AIExperts live - Improving user adoption with AI
Experts live - Improving user adoption with AI
 
DSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelinesDSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelines
 
DATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docx
DATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docxDATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docx
DATA COMMS-NETWORKS YR2 lecture 08 NAT & CLOUD.docx
 
原版一比一多伦多大学毕业证(UofT毕业证书)如何办理
原版一比一多伦多大学毕业证(UofT毕业证书)如何办理原版一比一多伦多大学毕业证(UofT毕业证书)如何办理
原版一比一多伦多大学毕业证(UofT毕业证书)如何办理
 
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
 
一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理
一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理
一比一原版(UCSF文凭证书)旧金山分校毕业证如何办理
 
一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理
 
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
 
Monthly Management report for the Month of May 2024
Monthly Management report for the Month of May 2024Monthly Management report for the Month of May 2024
Monthly Management report for the Month of May 2024
 
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
 
一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理
一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理
一比一原版兰加拉学院毕业证(Langara毕业证书)学历如何办理
 
Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...
 
UofT毕业证如何办理
UofT毕业证如何办理UofT毕业证如何办理
UofT毕业证如何办理
 
一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理
一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理
一比一原版(Unimelb毕业证书)墨尔本大学毕业证如何办理
 
一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理
一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理
一比一原版(GWU,GW文凭证书)乔治·华盛顿大学毕业证如何办理
 

Elementary statistical inference1

  • 1. Elementary Statistical Inference 1 석 진 환 GROOT SEMINAR
  • 2. Reference • Introduction to Mathematical Statistics, Hogg, 7e
  • 3. Which is better: machine learning or statistics? • Prediction vs. Explanation • Machine learning models are designed to make the most accurate predictions possible. • Statistical models are designed for inference about the relationships between variables. • Many statistical models can make predictions, but predictive accuracy is not their strength. Likewise, machine learning models provide various degrees of interpretability
  • 4. 목차 1. Estimation 1) Point estimation 2) Interval estimation 3) Order Statistics 2. Hypothesis Testing 3. The Method of Monte Carlo & Bootstrap Procedure
  • 5. Sampling and Statistics • We have a random variable 𝑋 of interest, but its pdf 𝑓 𝑥 or pmf 𝑝(𝑥) is not known 1. 𝑓 𝑥 or 𝑝(𝑥) is completely unknown 2. The form of 𝑓 𝑥 or 𝑝(𝑥) is know, down to a parameter 𝜃 Our information about the unknown distribution or the unknown parameters of the distribution of 𝑋 comes from a sample on 𝑋. 1. Sampling and Statistics
  • 6. • We consider the second classification(parametric inference) • 𝑋 has an exponential distribution, 𝐸𝑥𝑝 𝜃 , 𝜃 is unknown. • 𝑋 has a binomial distribution, b 𝑛, 𝑝 where 𝑛 is known, but 𝑝 is unknown. • 𝑋 has a gamma distribution, Γ 𝛼, 𝛽 , 𝛼 and 𝛽 are unknown. • 𝑓 𝑥 ; 𝜃 , 𝑝 𝑥 ; 𝜃 𝜃 ∈ Ω 1. Sampling and Statistics
  • 7. • The sample observations have the same distribution as 𝑋, and we denote them as the random variables 𝑋1, 𝑋2, … , 𝑋 𝑛, where 𝑛 denotes the sample size. • When the sample is actually drawn, we use lower case letters 𝑥1, 𝑥2, … , 𝑥 𝑛 as realizations of the sample. • Assume that 𝑋1, 𝑋2, … , 𝑋 𝑛 are mutually independent. • Definition 1 • If 𝑋1, 𝑋2, … , 𝑋 𝑛 iid, then these random variables constitute a random sample of size n from the common distribution. • vs Stochastic Process ?? 1. Sampling and Statistics
  • 8. • Definition 2 (Statistic) • 𝑋1, 𝑋2, … , 𝑋 𝑛 sample on a random variable 𝑋. Let 𝑇 = 𝑇(𝑋1, 𝑋2, … , 𝑋 𝑛) be a function of the sample. Then 𝑇 is called a statistic(통계량) • Once the sample is actually drawn, then 𝑡 is called the realization of 𝑇, where 𝑡 = 𝑇(𝑥1, 𝑥2, … , 𝑥 𝑛) • Example : 𝑋 = 𝑋1+𝑋2+⋯+𝑋 𝑛 𝑛 , 𝑆2 = 1 𝑛−1 𝑋𝑖 − 𝑋 2 are statistics • It makes sense to consider a statistic 𝑇, which is an estimator of 𝜃. • While we call 𝑇 an estimator of 𝜃, we call its realization 𝑡 an estimate of 𝜃. 1. Sampling and Statistics
  • 9. Estimation • Point estimation • 표본평균이 15니까, 모평균도 15일것이다! • MLE • Interval estimation • 표본평균이 15니까, 모평균은 (12.3, 14.7) 구간에 있을 것이다! • Confidence Interval 1. Sampling and Statistics
  • 10. Several Properties of point estimators • Unbiasedness(비편향성), Consistency(일치성), efficiency(효율성) • Definition 3 (Unbiasedness) • 𝑋1, 𝑋2, … , 𝑋 𝑛 : sample on a random variable 𝑋 with pdf 𝑓 𝑥 ; 𝜃 , 𝜃 ∈ Ω. Let 𝑇 = 𝑇(𝑋1, 𝑋2, … , 𝑋 𝑛) be a statistic. We say that 𝑇 is unbiased estimator of 𝜃 if 𝐸 𝑇 = 𝜃. 1. Sampling and Statistics
  • 11. Example : Maximum Likelihood Estimator(MLE) • 𝐿 𝜃 = 𝑖=1 𝑛 𝑓(𝑥𝑖, 𝜃), likelihood function of the random sample. • Often-used estimate is that value of 𝜃 which provides a maximum of 𝐿(𝜃). If it is unique, this is called the maximum likelihood estimator, and denote it as 𝜃, i.e., 𝜃 = 𝐴𝑟𝑔𝑚𝑎𝑥 𝜃 𝐿(𝜃). 1. Sampling and Statistics
  • 12. MLE with EXPONENTIAL DISTRIBUTION • Random sample 𝑋1, 𝑋2, … , 𝑋 𝑛 is the Γ(1, 𝜃) density. The log of likelihood function is 𝑙 𝜃 = 𝑙𝑜𝑔 𝑖=1 𝑛 1 𝜃 𝑒 − 𝑥 𝑖 𝜃 = −𝑛log𝜃 − 𝜃−1 𝑖=1 𝑛 𝑥𝑖 • 𝜕𝑙 𝜃 𝜕𝜃 = 0, 𝜕2 𝑙 𝜃 𝜕𝜃2 < 0 yields that 𝜃 = 𝑋 is the mle of 𝜃. • Because 𝐸 𝑋 = 𝜃, we have that 𝐸 𝑋 = 𝜃 and, hence, 𝜃 is an unbiased estimator of 𝜃. 1. Sampling and Statistics
  • 13. MLE with NORMAL DISTRIBUTION • 𝑋 ~ 𝑁 𝜇, 𝜎2 ∗ 𝜽 = 𝜇, 𝜎 • 𝑓 𝑥; 𝜽 = 1 2𝜋𝜎 𝑒 − 𝑥−𝜇 2 2𝜎2 • If 𝑋1, 𝑋2, … , 𝑋 𝑛 is a random variable on 𝑋, then 𝐿 𝜽 = 𝑖=1 𝑛 𝑓 𝑥𝑖; 𝜽 𝑙 𝜽 = log 𝐿 𝜽 = 𝑖=1 𝑛 log(𝑓(𝑥𝑖; 𝜃)) = − 𝑛 2 log 2𝜋 − 𝑛 log 𝜎 − 1 2 𝑖=1 𝑛 𝑥𝑖 − 𝜇 𝜎 2 𝜕𝑙 𝜕𝜇 = 𝜕𝑙 𝜕𝜎 = 0 yields 𝜇 = 𝑋 𝑎𝑛𝑑 𝜎2 = 𝑛−1 𝑖=1 𝑛 𝑋𝑖 − 𝑋 2 . • 𝜇 is an unbiased estimator of 𝜇 • 𝜎2 is a biased estimator of 𝜎2 , though converges to 𝜎2 as → ∞. 1. Sampling and Statistics
  • 14. 예제1(점추정) 베스킨라빈스 파인트의 무게는 정규분포를 따른다고 가정한다고 알고 있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다. MLE에 의한 파인트 무게의 모평균의 추정값은?? 표본 평균 : 321g -> 모평균의 추정값이 된다! 만약 회사 홈페이지에 310g이라고 써져있다면? 가설검정 1 2 3 4 5 6 7 8 9 10 310 315 330 320 325 325 310 315 320 340 2. Confidence Intervals
  • 15. Confidence Intervals • We discussed estimating 𝜃 by a statistic 𝜃 = 𝜃 𝑋1, 𝑋2, … , 𝑋 𝑛 , where 𝑋1, 𝑋2, … , 𝑋 𝑛 is a sample from the distribution of 𝑋. • When the sample is drawn, it is unlikely that the value of 𝜃 is the true value of the parameter. • Error must exist. • Embody estimate of error in terms of a confidence interval 2. Confidence Intervals
  • 16. Confidence Intervals • Definition 4 (Confidence Interval) • Let 𝑋1, 𝑋2, … , 𝑋 𝑛 be a sample on a random variable 𝑋, where 𝑋 has pdf 𝑓 𝑥: 𝜃 , 𝜃 ∈ Ω. Let 𝛼 be specified. Let 𝐿 = 𝐿 𝑋1, 𝑋2, … , 𝑋 𝑛 and 𝑈 = 𝑈 𝑋1, 𝑋2, … , 𝑋 𝑛 be two statistics. • We say that the interval (𝐿, 𝑈) is a 1 − 𝛼 100% confidence interval(신뢰 구간) for 𝜃 if 1 − 𝛼 = 𝑃 𝜃[𝜃 ∈ 𝐿, 𝑈 ] 1 − 𝛼 is called the confidence coefficient(신뢰도) of the interval Once the sample is drawn, the realized value of the confidence interval is 𝑙, 𝑢 = (𝐿 𝑥1, 𝑥2, … , 𝑥 𝑛 , 𝑈(𝑥1, 𝑥2, … , 𝑥 𝑛)) 2. Confidence Intervals
  • 17. 예제2(구간추정) 베스킨라빈스 파인트의 무게는 정규분포를 따른다고 알고있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다. 베스킨라빈스 파인트 무게의 평균에 대한 95% 신뢰구간은? 베스킨라빈스 파인트 무게의 모평균이 얼마인지는 모르겠지만, 표본을 보니, 95% 확률로 00~00에 포함되어 있다. 1 2 3 4 5 6 7 8 9 10 310 315 330 320 325 325 310 315 320 340 2. Confidence Intervals
  • 18. Confidence Interval for 𝜇 Under Normality • Suppose 𝑋1, 𝑋2, … , 𝑋 𝑛 are a random sample from a 𝑁 𝜇, 𝜎2 . Let 𝑋 and 𝑆2 denote the sample mean and sample variance, respectively. 𝑋 is mle of 𝜇 & 𝑛−1 𝑛 𝑆2 is mle of 𝜎2 𝑇 = 𝑋−𝜇 𝑆 𝑛 has a 𝑡-distribution with 𝑛 − 1 degrees of freedom. 1 − 𝛼 = 𝑃(−𝑡 𝛼 2,𝑛−1 < 𝑇 < 𝑡 𝛼 2,𝑛−1) = 𝑃(−𝑡 𝛼 2,𝑛−1 < 𝑋−𝜇 𝑆 𝑛 < 𝑡 𝛼 2,𝑛−1) = 𝑃( 𝑋 − 𝑡 𝛼 2,𝑛−1 𝑆 𝑛 < 𝜇 < 𝑋 + 𝑡 𝛼 2,𝑛−1 𝑆 𝑛 ) ∴ 𝐿 𝑋1, 𝑋2 , … , 𝑋 𝑛 = 𝑋 − 𝑡 𝛼 2,𝑛−1 𝑆 𝑛 and U 𝑋1, 𝑋2 , … , 𝑋 𝑛 = 𝑋 + 𝑡 𝛼 2,𝑛−1 𝑆 𝑛 2. Confidence Intervals
  • 19. 𝑡-interval and standard error • Once the sample is drawn, let 𝑥 and 𝑠 denote the realized values of the statistics 𝑋 and 𝑆. Then 1 − 𝛼 100% conficence interval for 𝜇 is given by ( 𝑥 − 𝑡 𝛼 2,𝑛−1 𝑠 𝑛 , 𝑥 + 𝑡 𝛼 2,𝑛−1 𝑠 𝑛 ) t-interval Standard error 2. Confidence Intervals
  • 20. 예제2 풀이 1 2 3 4 5 6 7 8 9 10 310 315 330 320 325 325 310 315 320 340 𝑥 = 321, 𝑠 = 11.79 𝑛 = 10, 𝑡 𝛼 2,𝑛−1 = 2.262 95% 신뢰구간 = (312.5665, 329.4335) 2. Confidence Intervals
  • 21. Central Limit Theorem(CLT) • The last example depends on the normality of the sampled items • However, this is approximately true even if the sampled items are not drawn from a normal distribution. • Theorem 1 (Central Limit Theorem) • Let 𝑋1, 𝑋2, … , 𝑋 𝑛 is a random sample with mean 𝜇 and finite variance 𝜎2 . Define a random variable 𝑍 𝑛 = 𝑋−𝜇 𝜎 𝑛 . Then 𝑍 𝑛 → 𝐷 𝑍, where 𝑍~𝑁(0,1). • 𝑋를 사용하는 경우 • 𝑛 이 충분히 큰 경우, 𝑡-분포를 사용할까? 정규분포를 사용할까? -> 𝑡-분포 : more conservative 2. Confidence Intervals
  • 22. Confidence Intervals for Difference in Means • We compare the means of 𝑋 and 𝑌, denoted by 𝜇1 and 𝜇2, respectively. Assume that the variances of 𝑋 and Y are finite and denote them as 𝜎1 2 and 𝜎2 2 , respectively. • Goal : Estimate ∆= 𝜇1 − 𝜇2. • What is confidence intervals of ∆= 𝜇1 − 𝜇2? • Sampling : 𝑋1, 𝑋2, … , 𝑋 𝑛1 & 𝑌1, 𝑌2, … , 𝑌𝑛2 • Let ∆ = 𝑋 − 𝑌. The statistic ∆ is unbiased estimator of ∆. • 𝑉𝑎𝑟 ∆ = 𝜎1 2 𝑛1 + 𝜎2 2 𝑛2 . By CLT, 𝑍 = ∆−∆ 𝑆1 2 𝑛1 + 𝑆2 2 𝑛2 ~𝑁(0,12) ( 𝑥 − 𝑦 − 𝑧 𝛼 2 𝑠1 2 𝑛1 + 𝑠2 2 𝑛2 , 𝑥 − 𝑦 + 𝑧 𝛼 2 𝑠1 2 𝑛1 + 𝑠2 2 𝑛2 ) 2. Confidence Intervals
  • 23. Confidence Intervals for Difference in Means • 𝑋 and 𝑌 are normal with the same variance; i.e., 𝜎2 = 𝜎1 2 = 𝜎2 2 . • Estimator 𝑆 𝑝 2 of 𝜎2 is weighted average of 𝑆1 2 and 𝑆2 2 . → pooled estimator → 𝑆 𝑝 2 = 𝑛1−1 𝑆1 2+ 𝑛2−1 𝑆2 2 𝑛1+𝑛2−2 • 𝑇 = 𝑋− 𝑌 − 𝜇1−𝜇2 𝜎 𝑛1 −1+𝑛2 −1 𝑛−2 𝑆 𝑝 2 𝑛−2 𝜎2 has a 𝑡-distribution with 𝑛 − 2 degrees of freedom. • 1 − 𝛼 100% confidence interval for ∆= 𝜇1 − 𝜇2: • (( 𝑥- 𝑦) − 𝑡 𝛼 2,𝑛−2 𝑠 𝑝 1 𝑛1 + 1 𝑛2 , 𝑥− 𝑦 + 𝑡 𝛼 2,𝑛−2 𝑠 𝑝 1 𝑛1 + 1 𝑛2 )) 2. Confidence Intervals
  • 24. Order Statistics(순서통계량) • 전구가 100개 있고, 𝑖 번째 전구의 수명을 확률변수 𝑋𝑖라고 하자. 첫 번째 꺼진 전구의 수명을 𝑌1, 두 번째 꺼진 전구의 수명을 𝑌2, … , 100번째 꺼진 전구의 수명을 𝑌100 이라고 하자. • 이렇게 순서대로 얻어진 확률변수를 순서통계량이라고 한다. • 𝑋1, 𝑋2, … , 𝑋 𝑛이 독립이어도, 𝑌1, 𝑌2, … , 𝑌𝑛은 독립이 아닐 수 있다. 3. Order Statistics
  • 25. Order Statistics(순서통계량) • Theorem 2 • 𝑋 : continuous random variable whose pdf 𝑓(𝑥) has support 𝒮 = 𝑎, 𝑏 , where −∞ ≤ 𝑎 < 𝑏 ≤ ∞. Let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 denote the 𝑛 order statistics 𝑋1, 𝑋2, … , 𝑋 𝑛 from 𝑋. Then joint pdf of 𝑌1, 𝑌2, … , 𝑌𝑛 is given by 𝑔 𝑦1, 𝑦2, … , 𝑦𝑛 = 𝑛! 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓 𝑦𝑛 𝑎 < 𝑦1 < 𝑦2 < ⋯ < 𝑦𝑛 < 𝑏 0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒 3. Order Statistics
  • 26. • Proof • 𝑔 𝑦1, 𝑦2, … , 𝑦𝑛 = 𝑖=1 𝑛! 𝒥𝑖 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓(𝑦𝑛) = 𝑛! 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓 𝑦𝑛 𝑎 < 𝑦1 < 𝑦2 < ⋯ < 𝑦𝑛 < 𝑏 0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒 ∵ 𝒥𝑖 = 1 • Important statistics using order statistics • Range = 𝑌𝑛 − 𝑌1 • Midrange = Y1+𝑌𝑛 2 • Median = 𝑌 𝑛+1 2 , 𝑛 𝑖𝑠 𝑜𝑑𝑑. 3. Order Statistics
  • 27. Quantiles • Let 𝑋 be a random variable with a continuous cdf 𝐹(𝑥). For 0 < 𝑝 < 1, define the 𝒑th quantile of 𝑋 to be 𝜉 𝑝 = 𝐹−1 (𝑝). • e.g) 𝜉0.5 is median of 𝑋. • What is estimator of 𝜉 𝑝? • Confidence Interval for 𝜉 𝑝? 3. Order Statistics
  • 28. • Let 𝑋1, 𝑋2, … , 𝑋 𝑛 be a random sample from the distribution of 𝑋 and let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 be the corresponding order statistics. Let 𝑘 be the greatest integer less than or equal to 𝑝(𝑛 + 1). • 𝐸 𝐹 𝑌𝑘 = 𝑎 𝑏 𝐹 𝑦 𝑘 𝑔 𝑘 𝑦 𝑘 𝑑𝑦 𝑘 = 0 1 𝑛! 𝑘−1 ! 𝑛−𝑘 ! 𝑧 𝑘 1 − 𝑧 𝑛−𝑘 𝑑𝑧 = 𝑘 𝑛+1 ≈ 𝑝 • We take 𝑌𝑘 as an estimator of the quantile 𝜉 𝑝. • 𝑌𝑘 is called the 𝒑th sample quantile(or, 100𝒑th percentile of the sample) • Sample quantiles are useful descriptive statistics. 3. Order Statistics
  • 29. Application of Quantile : Five-number summary • A five-number summary of the data consist of the following sample quantiles • The minimum(𝑌1), the first quartile(𝑌.25(𝑛+1) = 𝜉.25), the median(𝑌.50(𝑛+1) = 𝜉.50), the third quartile(𝑌.75(𝑛+1) = 𝜉.75), and the maximum(𝑌𝑛) • Use the notation 𝑄1, 𝑄2 , and 𝑄3 to denote, respectively, the first quartile, median, and third quartile of the sample. 3. Order Statistics
  • 30. 예제2 • 확률변수 𝑋의 샘플 15개 데이터가 다음과 같다. • The minimum : 𝑦1 = 56 • 𝑄1 = 𝑦4 = 94 • 𝑄2 = 𝑦8 = 102 • 𝑄3 = 𝑦12 = 108 • The maximum : 𝑦15 = 116 56 70 89 94 96 101 102 102 102 105 106 108 110 113 116 3. Order Statistics
  • 31. Application of Quantile : Box-Whisker plot • The five-number summary is the basis for a useful and quick plot of the data. • ℎ = 1.5(𝑄3 − 𝑄1) • 𝐥𝐨𝐰𝐞𝐫 𝐟𝐞𝐧𝐜𝐞 𝐿𝐹 = 𝑄1 − ℎ, 𝐮𝐩𝐩𝐞𝐫 𝐟𝐞𝐧𝐜𝐞(𝑈𝐹) = 𝑄3 + ℎ • Points that lie outside the fences are called potential outliers and they are denoted by ‘o’ on the boxplot. • The whiskers protrude from the sides of the box to what are called adjacent points, which are the points within the fences but closest to the fences. 3. Order Statistics
  • 32. 56 70 89 94 96 101 102 102 102 105 106 108 110 113 116 The minimum : 𝑦1 = 56 𝑄1 = 𝑦4 = 94 𝑄2 = 𝑦8 = 102 𝑄3 = 𝑦12 = 108 The maximum : 𝑦15 = 116 3. Order Statistics
  • 33. Application of Quantile : Diagnostic plot • In practice, we often assume that the data follow a certain distribution. Such an assumption needs to be checked and there are many statistical tests which do so Goodness of fit(위키피디아) • We discuss one such diagnostic plot in this regard. • q-q plot 3. Order Statistics
  • 34. q-q plot • Suppose 𝑋 is a random variable with cdf 𝐹 𝑥 − 𝑎 𝑏 , where 𝐹(𝑥) is known but 𝑎 and 𝑏 > 0 may not be. • Let 𝑍 = (𝑋 − 𝑎) 𝑏. Then, 𝑍 has cdf 𝐹(𝑧). Let 0 < 𝑝 < 1 and let 𝜉 𝑋,𝑝 be the 𝑝th quantile of 𝑋 and 𝜉 𝑍,𝑝 be the 𝑝th quantile of 𝑍. • 𝑝 = 𝑃 𝑋 ≤ 𝜉 𝑋,𝑝 = 𝑃[𝑍 ≤ 𝜉 𝑋,𝑝−𝑎 𝑏 ] ⇒ 𝜉 𝑋,𝑝 = 𝑏𝜉 𝑍,𝑝 + 𝑎 • Thus, the quantiles of 𝑋 are linearly related to the quantiles of 𝑍. 3. Order Statistics
  • 35. q-q plot • 𝑋1, 𝑋2, … , 𝑋 𝑛 : Random sample from 𝑋 ⇒ 𝑌1 < ⋯ < 𝑌𝑛 the order statistics. • For 𝑘 = 1, … , 𝑛, let 𝑝 𝑘 = 𝑘 𝑛+1 . ⇒ 𝑌𝑘 is an estimator of 𝜉 𝑋,𝑝. • Denote the corresponding quantiles of the cdf 𝐹 𝑧 by 𝜉 𝑍,𝑝 = 𝐹−1(𝑝 𝑘) • The plot of 𝑌𝑘 versus 𝜉 𝑍,𝑝 is called a q-q plot. 3. Order Statistics
  • 36. • Figure 4.1 contain q-q plots of three different Distributions • Panel C appears to be most linear. We may assume that the data were generated from a Laplace distribution 3. Order Statistics
  • 37. Confidence Intervals for Quantiles • For a sample size 𝑛 on 𝑋, let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 be the order statistics. Let 𝑘 = [ 𝑛 + 1 𝑝]. Then the 100𝑝th sample percentile 𝑌𝑘 is a point estimator of 𝜉 𝑝. • We now derive distribution free confidence interval for 𝜉 𝑝. • Let 𝑖 < 𝑛 + 1 𝑝 < 𝑗, and consider the order statistics 𝑌𝑖 < 𝑌𝑗 and the event 𝑌𝑖 < 𝜉 𝑝 ∩ 𝑌𝑗 > 𝜉 𝑝 = [적어도 i개는 𝜉 𝑝보다 작고 𝑗번째는 𝜉 𝑝보다 큰 경우] • The probability of success is 𝑃 𝑋 < 𝜉 𝑝 = 𝐹 𝜉 𝑝 = 𝑝. 𝑃 𝑌𝑖 < 𝜉 𝑝 < 𝑌𝑗 = 𝑤=𝑖 𝑗−1 𝑛 𝑤 𝑝 𝑤 1 − 𝑝 𝑛−𝑤 (= 0.95 → 𝑌𝑖, 𝑌𝑗 ) 3. Order Statistics
  • 38. Example: Confidence Interval for the Median. • 𝜉.50 ∶ median of 𝐹(𝑥), 𝑄2 ∶ sample median − point estimator 𝑜𝑓𝜉.50 • Select 0 < 𝛼 < 1 and Take 𝑐 𝛼 2 to be the 𝛼 2 𝑡ℎ quantile of a binomial S ~ 𝑏(𝑛, 1 2 ). • 𝑃 𝑆 ≤ 𝑐 𝛼 2 = 𝑃 𝑆 ≥ 𝑛 − 𝑐 𝛼 2 = 𝛼 2. • Thus it follows that 𝑃 𝑌𝑐 𝛼 2+1 < 𝜉.50 < 𝑌𝑛−𝑐 𝛼 2 = 1 − 𝛼 • Hence, when the sample is drawn, (𝑦𝑐 𝛼 2+1, 𝑦𝑛−𝑐 𝛼 2 ) is a (1 − 𝛼)100% confidence interval for 𝜉.50. 3. Order Statistics
  • 39. 예제3 • 모기 4마리를 채집해서 모기 4마리의 수명을 관찰했다. 모기의 수명이 평균이 1인 지수분포를 따를 때, 제일 오래 산 모기의 수 명이 3 이상일 확률은? • Let 𝑌1 < 𝑌2 < 𝑌3 < 𝑌4 be the order statistics of a random sample of size 4 from the distribution having pdf 𝑓 𝑥 = 𝑒−𝑥, 0 < 𝑥 < ∞. Find 𝑃 3 ≤ 𝑌4 . • 𝑃 𝑌4 ≤ 𝑡 = 1 − 𝑒−𝑡 4 ∴ 𝑃 𝑌4 ≥ 3 = 1 − 1 − 𝑒−3 4 = 0.1848 3. Order Statistics
  • 40. 예제4 • 모기 4마리를 채집해서 모기 4마리의 수명을 관찰했다. 모기의 수명이 0에서 1의 값을 가지는 균등분포를 따를 때, 제일 오래 산 모기와 제일 빨리 죽은 모기의 수명의 차이가 0.5 미만일 확 률은? • Let 𝑌1 < 𝑌2 < 𝑌3 < 𝑌4 be the order statistics of a random sample of size 4 from the distribution having pdf 𝑓 𝑥 = 1, 0 < 𝑥 < 1. Find the probability that the range of the random sample is less then 1 2 . 3. Order Statistics
  • 41. • 𝑠𝑜𝑙) 𝑃 𝑌4 − 𝑌1 > 1 2 = 0 1 2 𝑦1+ 1 2 1 𝑔1,4(𝑦1, 𝑦4)𝑑𝑦4 𝑑𝑦1 • 𝑔1,4 𝑦1, 𝑦4 = 𝑦1 𝑦4 𝑦2 𝑦4 𝑔 𝑦1, 𝑦2, 𝑦3, 𝑦4 𝑑𝑦3 𝑑𝑦2 = 12 𝑦4 − 𝑦1 2 • 𝑃 𝑌4 − 𝑌1 ≤ 1 2 = 1 − 0 1 2 𝑦1+ 1 2 1 12 𝑦4 − 𝑦1 2 𝑑𝑦4 𝑑𝑦1 = 11 16 3. Order Statistics
  • 42. Introduction to Hypothesis Testing 4 Introduction to Hypothesis Testing Statistical Inference Estimation Hypothesis Testing Point estimation Interval estimation
  • 43. 예제5(점추정/가설검정) 베스킨라빈스 파인트의 무게는 정규분포를 따른다고 가정한다. 베스킨라빈스에 10 번을 가서, 무게를 재봤더니 다음과 같았다. MLE에 의한 파인트 무게의 모평균의 추정값은?? 표본 평균 : 321g -> 모평균의 추정값이 된다! 만약 회사 홈페이지에 310g이라고 써져있다면? 가설검정 1 2 3 4 5 6 7 8 9 10 310 315 330 320 325 325 310 315 320 340 4 Introduction to Hypothesis Testing
  • 44. • 귀무가설 : 파인트 무게는 평균 310g이다. • 연구가설 : 파인트 무게는 평균 310g이 아니다. • 모평균의 추정값 : 321g -> 귀무가설 기각 (X) • 기각역 설정 -> 추정값이 기각역에 포함??(O) 1 2 3 4 5 6 7 8 9 10 평균 310 315 330 320 325 325 310 315 320 340 321 4 Introduction to Hypothesis Testing
  • 45. 모분포 : 평균 100, 분산 30 Sampling 표본 평균 : 90 50개 표본 평균 : 80 표본 평균 : 40 표본 평균 : 10 표본 평균 : 1 표본 평균 : 150 귀무가설 귀무가설의 모평균에서 멀어질수록 귀무가설이 틀렸다는 생각이 커진다. WHY? 50개 뽑았으면 우연이 아니라 필연! 4 Introduction to Hypothesis Testing
  • 46. Hypothesis • 𝐻0 : Null hypothesis(귀무가설) • Represents no change or no difference from the past • 𝐻1 : Alternative hypothesis(대립가설/연구가설) • Represents change or difference • Research worker’s hypothesis 𝐻0 ∶ 𝜃 ∈ 𝜔0 versus 𝐻1 ∶ 𝜃 ∈ 𝜔1 (where 𝜔0 ∪ 𝜔1 = Ω , and 𝜔0 ∩ 𝜔1 = ∅) • The decision rule to take 𝐻0 or 𝐻0 is based on a sample 𝑋1, 𝑋2, … , 𝑋 𝑛 from 𝑋 • Hence, the decision could wrong. 4 Introduction to Hypothesis Testing
  • 47. Decision Rule • 𝑋1, … , 𝑋 𝑛 : random sample from the distribution of 𝑋. • Consider testing the hypotheses : • 𝐻0 ∶ 𝜃 ∈ 𝜔0 versus 𝐻1 ∶ 𝜃 ∈ 𝜔1 • Denote the space of the sample by 𝒟 = space{(𝑋1, … , 𝑋 𝑛)}. • Definition 5 (Statistical Test, Critical Region) • A test of 𝐻0 versus 𝐻1 is based on a subset 𝐶 of 𝒟. • This set 𝐶 is called the critical region(기각역) and its corresponding decision rule is: Reject 𝐻0 if 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 Retain 𝐻0 if 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 𝑐 4 Introduction to Hypothesis Testing
  • 48. • the decision could wrong. • Type 1, Type 2 error • Goal : Select a critical region from all possible critical regions which minimizes the probabilities of these two errors. • in general, not possible (trade-off) • Procedure • Select critical regions which bound the probability of Type 1 error • Among these critical regions we try to select one which minimizes Type 2 error. 4 Introduction to Hypothesis Testing
  • 49. • Definition 5 (Critical region of size 𝛼) • We say a critical region 𝐶 is of size 𝛼 if 𝛼 = max 𝜃∈𝜔0 𝑃 𝜃[ 𝑋1, … 𝑋 𝑛 ∈ 𝐶] • Over all critical region of size 𝛼, we want to consider critical regions which have lower probabilities of Type 2 error. • 𝑃 𝜃 Type 2 Error = 𝑃 𝜃 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 𝑐 , 𝜃 ∈ 𝜔1 • we desire to maximize 𝑃 𝜃[ 𝑋1, … , 𝑋 𝑛 ∈ 𝐶]. (power of the test at 𝜃) • Definition 6 (Power function) • Power function of a critical region to be 𝛾 𝐶 𝜃 = 𝑃 𝜃 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 ; 𝜃 ∈ 𝜔1 4 Introduction to Hypothesis Testing
  • 50. Better critical region • Given two critical regions 𝐶1 and 𝐶2 which are both of size 𝛼, 𝐶1 is better than 𝐶2 if 𝛾 𝐶1 𝜃 ≥ 𝛾 𝐶2 (𝜃) for all 𝜃 ∈ 𝜔1. • Type1 에러의 최대값이 𝛼로 같으면, Type2 에러가 작은 것이 더 좋은 기각역이다. 4 Introduction to Hypothesis Testing
  • 51. Test for a Binomial Proportion of Success • 𝑋~𝐵𝑒𝑟(𝑝), 𝑝0 : probability of dying with some standard treatment • We want to test at size 𝛼, 𝐻0 ∶ 𝑝 = 𝑝0 versus 𝐻1 ∶ 𝑝 < 𝑝0 • Let 𝑋1, … , 𝑋 𝑛 be a random sample from the distribution of X and let 𝑆 = 𝑖=1 𝑛 𝑋𝑖 . • An intuitive decision rule is : Reject 𝐻0 in favor of 𝐻1 if 𝑆 ≤ 𝑘, where 𝑘 is such that 𝛼 = 𝑃 𝐻0 [𝑆 ≤ 𝑘]. • For example, suppose 𝑛 = 20, 𝑝0 = 0.7, and 𝛼=0.15 • 𝑃 𝐻0 𝑆 ≤ 11 = 0.1133 and 𝑃 𝐻0 𝑆 ≤ 12 = 0.2277 ⋯ 𝑘 = 11? 12? • On the conservative side, choose 𝑘 to be 11 and 𝛼 = 0.1133. • 𝛾 𝑝 = 𝑃𝑝 𝑆 ≤ 𝑘 , 𝑝 < 𝑝0 4 Introduction to Hypothesis Testing
  • 52. 4 Introduction to Hypothesis Testing
  • 53. Nomenclature • 𝐻0 ∶ 𝑝 = 𝑝0 completely specifies the underlying distribution, it is called a simple hypothesis. ( 𝜔0 = 1) • 𝐻1 ∶ 𝑝 < 𝑝0 is composite hypothesis. • Frequently, 𝛼 is also called the significance level(유의수준) of the test • Or, ”maximum of probabilities of committing an error of Type 1 error” • Or, “maximum of the power of the test when 𝐻0 is true” 4 Introduction to Hypothesis Testing
  • 54. 𝒕-test : Test for 𝜇 Under Normality • Let 𝑋 have a 𝑁(𝜇, 𝜎2) distribution. Consider the hypotheses 𝐻0 ∶ 𝜇 = 𝜇0 𝑣𝑒𝑟𝑠𝑢𝑠 𝐻1 ∶ 𝜇 > 𝜇0 • Assume that the desired size of the test is 𝛼, for 0 < 𝛼 < 1. • Our intuitive rejection rule is to reject 𝐻0 if 𝑋 is much larger than 𝜇0. • 𝑇 = 𝑋−𝜇0 𝑆 𝑛 , 𝑡-distribution with 𝑛-1 degree of freedom. It follows that this rejection rule has exact level 𝛼 : Reject 𝐻0 if 𝑇 = 𝑋 − 𝜇0 𝑆 𝑛 ≥ 𝑡 𝛼,𝑛−1, where 𝛼 = 𝑃[𝑇 > 𝑡 𝛼,𝑛−1] 4 Introduction to Hypothesis Testing
  • 55. • Large sample rule ?? • In practice, we may not be willing to assume that the population is normal. Usually 𝑡-critical values are larger than 𝑧-critical values • Hence, the 𝑡-test is conservative relative to the large sample test. So, in practice, many statisticians often use the 𝑡-test. 4 Introduction to Hypothesis Testing
  • 56. 예제6(가설검정 𝑡-test) 베스킨라빈스 파인트의 무게는 모평균이 310g 인 정규분포를 따 른다고 알고있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다. 가설 설정 : • 귀무가설(𝐻0) : 𝜇 = 310 • 연구가설(𝐻1) : 𝜇 > 310 1 2 3 4 5 6 7 8 9 10 평균 310 315 330 320 325 325 310 315 320 340 321 4 Introduction to Hypothesis Testing
  • 57. • 기각역 설정(유의수준(𝛼) = 5%) • Intuitive rejection rule : • 𝑋 much larger than 310 or much smaller than 310 • 𝑇 = 𝑋−310 𝑆 10 ≥ 𝑡0.05,9 • Critical region of size 0.05 • 𝐶 = { 𝑥 ≥ 312.1170 } • 𝑥 = 321 -> 귀묵가설 기각 • Conclusion • 베스킨라빈스 파인트 모평균은 310g이 아니고, 310g 보다 크다. 4 Introduction to Hypothesis Testing
  • 58. Randomized tests and 𝑝-value • Let 𝑋1, … , 𝑋10 be random sample of size 𝑛=10 from a Poisson distribution with mean 𝜃. • A critical region for testing 𝐻0 ∶ 𝜃 = 0.1 against 𝐻1 ∶ 𝜃 > 0.1 is given by 𝑌 = 𝑖=1 10 𝑋𝑖 ≥ 3. The statistic 𝑌 has a Poisson distribution with mean 10𝜃. • Significance level of the test is 𝑃 𝑌 ≥ 3 = 1 − 𝑃 𝑌 ≤ 2 = 1 − 0.920 = 0.080 • If the critical region defined by 𝑖=1 10 𝑋𝑖 ≥ 4 is used, the significance level is 𝛼 = 𝑃 𝑌 ≥ 4 = 1 − 𝑃 𝑌 ≤ 3 = 0.019 How we achieve a significance level of 𝛼 = 0.05?? 4 Introduction to Hypothesis Testing
  • 59. • Let 𝑊 have a Bernoulli distribution with probability of success equal to 𝑃 𝑊 = 1 = 0.050 − 0.019 0.080 − 0.019 = 31 61 • Assume that 𝑊 is selected independently of the sample. Consider the rejection rule Reject 𝐻0 if 1 10 𝑥𝑖 ≥ 4 or if 1 10 𝑥𝑖 = 3 and 𝑊 = 1. • The significance level of this rule is 𝑃 𝐻0 𝑌 ≥ 4 + 𝑃 𝑌 = 3 ∩ 𝑊 = 1 = 𝑃 𝐻0 𝑌 ≥ 4 + 𝑃 𝐻0 𝑌 = 3 𝑃 𝑊 = 1 = 0.019 + 0.061 31 61 = 0.05 • The process of performing the auxiliary experiment is referred to as a randomized test. 4 Introduction to Hypothesis Testing
  • 60. Observed Significance Level • Not many statisticians like randomized tests in practice, because the use of them means that two statisticians could make the same assumptions, observe the same data, and yet make different decisions. • As a matter of fact, many statisticians report what are commonly called observed significance level or 𝒑-values. 유의수준을 정했는데 그에 따른 기각역을 정할 수 없으면 그냥 p-value를 살피자. 4 Introduction to Hypothesis Testing
  • 61. 모분포 : 평균 100, 분산 30 Sampling 50개 표본 평균 : 30 표본 평균 : 180 귀무가설 두 표본이 얼마나 귀무가설을 벗어나는지 측정하고 싶다. 표본평균 30이 뽑히는 것만큼 극단적으로 관찰될 확률은? 표본평균 180이 뽑히는 것만큼 극단적으로 관찰될 확률은? 유의수준을 정해준 것은 귀무가설을 얼마나 벗어나면 기각을 시킬지 미리 정해준것! 4 Introduction to Hypothesis Testing P-value
  • 62. • 𝑝-value is the observed “tail” probability of a statistic being at least as extreme as the particular observed value when 𝐻0 is true. • If 𝑌 = 𝑢(𝑋1, 𝑋2, … , 𝑋 𝑛) is the statistic to be used in a test of 𝐻0 and if the critical region is of the form 𝑢 𝑥1, 𝑥2, … , 𝑥 𝑛 ≤ 𝑐, an observed value 𝑢 𝑥1, 𝑥2, … , 𝑥 𝑛 = 𝑑 would mean that the 𝑝-value = 𝑃(𝑌 ≤ 𝑑; 𝐻0) 4 Introduction to Hypothesis Testing
  • 63. 예제7 • 𝑋1, 𝑋2, … 𝑋25 a random sample from 𝑁(𝜇, 𝜎2 = 4). 𝑋~𝑁(𝜇, 0.16) • Test 𝐻0 ∶ 𝜇 = 77 against the one-sided alternative hypothesis 𝐻1 ∶ 𝜇 < 77 • 𝑥 = 76.1 • 𝑧 − 𝑠𝑐𝑜𝑟𝑒 : -2.25 • 𝑝 − 𝑣𝑎𝑙𝑢𝑒 ∶ ∅ −2.25 = 0.012 • If we were using a significance level of 𝛼 = 0.05, we would reject 𝐻0 and accept 𝐻1 ∶ 𝜇 < 77 because 0.012 < 0.05 4 Introduction to Hypothesis Testing

Editor's Notes

  1. Suppose “success” is dying from a certain disease and 𝑝 0 is the probability of dying with some standard treatment. A new treatment is used on several patients, and it is hoped that the probability of dying under this new treatment is less than 𝑝 0 .