1. This document discusses concepts related to statistical inference including point estimation, interval estimation, hypothesis testing, and Monte Carlo simulation.
2. It provides examples of how to calculate point estimators like the sample mean and maximum likelihood estimators for distributions like the normal and exponential.
3. Confidence intervals are defined as intervals that have a known probability of containing the true population parameter based on a sample. Formulas for 95% confidence intervals of the mean are given assuming normality and using the t-distribution.
📺Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 7: Estimating Parameters and Determining Sample Sizes
7.1: Estimating a Population Proportion
Statistical inference: Hypothesis Testing and t-testsEugene Yan Ziyou
This deck was used in the IDA facilitation of the John Hopkins' Data Science Specialization course for Statistical Inference. It covers the topics in week 3 (hypothesis testing and t tests).
The data and R script for the lab session can be found here: https://github.com/eugeneyan/Statistical-Inference
Get to know more about Directional and Non-Directional Hypothesis tests like one-tail, two-tailed along with 2 sample tests, paired difference T-test, if you are interested to implement the same in python check out my other blogs. Ping me @ google #bobrupakroy Happy Data Science Talk soon!
Learn to perform hypothesis testing with their stages of hypothesis testing with ease with the help of Excel and much more. See ya soon. Ping @ #bobrupakroy
This document discusses methods for estimating population parameters from sample data, including point estimation, bias, confidence intervals, sample size determination, and hypothesis testing. Key points include defining point estimates as single values representing plausible population values based on sample data, describing how to calculate confidence intervals for population proportions and means using z-tests and t-tests, and outlining how to determine necessary sample sizes to achieve a desired level of accuracy and confidence.
Multiple sample test - Anova, Chi-square, Test of association, Goodness of Fit Rupak Roy
Detailed demonstration of Multiple Sample Test like Analysis of Variance (ANOVA), kinds of ANOVA One Way, Two Way, Chi-square with their assumptions and applications using excel, and much more.
Let me know if anything is needed. Happy to help. ping @ #bobrupakroy
Types of Probability Distributions - Statistics IIRupak Roy
Get to know in detail the definitions of the types of probability distributions from binomial, poison, hypergeometric, negative binomial to continuous distribution like t-distribution and much more.
Let me know if anything is required. Ping me at google #bobrupakroy
[Karger+ NIPS11] Iterative Learning for Reliable Crowdsourcing SystemsShuyo Nakatani
This document discusses methods for improving the reliability of crowdsourced systems by identifying spam workers. It proposes an iterative algorithm that exchanges messages between tasks and workers to predict answers and estimate error rates. The algorithm guarantees an upper bound on error rates that decreases exponentially as the number of iterations increases, allowing highly accurate predictions even with some unreliable workers. Experimental results demonstrate the algorithm achieves lower error rates than other common methods.
📺Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 7: Estimating Parameters and Determining Sample Sizes
7.1: Estimating a Population Proportion
Statistical inference: Hypothesis Testing and t-testsEugene Yan Ziyou
This deck was used in the IDA facilitation of the John Hopkins' Data Science Specialization course for Statistical Inference. It covers the topics in week 3 (hypothesis testing and t tests).
The data and R script for the lab session can be found here: https://github.com/eugeneyan/Statistical-Inference
Get to know more about Directional and Non-Directional Hypothesis tests like one-tail, two-tailed along with 2 sample tests, paired difference T-test, if you are interested to implement the same in python check out my other blogs. Ping me @ google #bobrupakroy Happy Data Science Talk soon!
Learn to perform hypothesis testing with their stages of hypothesis testing with ease with the help of Excel and much more. See ya soon. Ping @ #bobrupakroy
This document discusses methods for estimating population parameters from sample data, including point estimation, bias, confidence intervals, sample size determination, and hypothesis testing. Key points include defining point estimates as single values representing plausible population values based on sample data, describing how to calculate confidence intervals for population proportions and means using z-tests and t-tests, and outlining how to determine necessary sample sizes to achieve a desired level of accuracy and confidence.
Multiple sample test - Anova, Chi-square, Test of association, Goodness of Fit Rupak Roy
Detailed demonstration of Multiple Sample Test like Analysis of Variance (ANOVA), kinds of ANOVA One Way, Two Way, Chi-square with their assumptions and applications using excel, and much more.
Let me know if anything is needed. Happy to help. ping @ #bobrupakroy
Types of Probability Distributions - Statistics IIRupak Roy
Get to know in detail the definitions of the types of probability distributions from binomial, poison, hypergeometric, negative binomial to continuous distribution like t-distribution and much more.
Let me know if anything is required. Ping me at google #bobrupakroy
[Karger+ NIPS11] Iterative Learning for Reliable Crowdsourcing SystemsShuyo Nakatani
This document discusses methods for improving the reliability of crowdsourced systems by identifying spam workers. It proposes an iterative algorithm that exchanges messages between tasks and workers to predict answers and estimate error rates. The algorithm guarantees an upper bound on error rates that decreases exponentially as the number of iterations increases, allowing highly accurate predictions even with some unreliable workers. Experimental results demonstrate the algorithm achieves lower error rates than other common methods.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 7: Estimating Parameters and Determining Sample Sizes
7.3: Estimating a Population Standard Deviation or Variance
The document discusses a one-sample t-test used to compare sample data to a standard value. It provides an example comparing intelligence scores of university students to the average score of 100. The sample of 6 students had a mean of 120. Running a one-tailed t-test in SPSS, the results showed the mean score was significantly higher than 100 with t(5)=3.15, p=.02. This allows the inference that the population mean intelligence at the university is greater than the standard score of 100.
This document provides an overview of statistical estimation and inference. It discusses point estimation, which provides a single value to estimate an unknown population parameter, and interval estimation, which gives a range of plausible values for the parameter. The key aspects of interval estimation are confidence intervals, which provide a probability statement about where the true population parameter lies. The document also covers important concepts like sampling distributions, the central limit theorem, and factors that influence the width of a confidence interval like sample size. Examples are provided to demonstrate calculating point estimates, confidence intervals, and dealing with independent samples.
This article provides a brief discussion on several statistical parameters that are most commonly used in any measurement and analysis process. There are a plethora of such parameters but the most important and widely used are briefed in here.
This document discusses hypothesis testing for claims about population proportions and the difference between two population proportions. It provides information on type I and type II errors. Examples are provided to demonstrate hypothesis testing for a single proportion claim and the difference between two proportions. The examples show setting up the null and alternative hypotheses, checking assumptions, calculating the test statistic, determining the p-value or comparing to the critical value, and making a conclusion. Confidence intervals are also discussed as a way to estimate population proportions and differences between proportions. The examples provide step-by-step workings to test claims about spending behaviors with different denominations of money.
Linear regression is an approach for modeling the relationship between one dependent variable and one or more independent variables.
Algorithms to minimize the error are
OLS (Ordinary Least Square)
Gradient Descent and much more.
Let me know if anything is required. Ping me at google #bobrupakroy
Statistical estimators are functions used to estimate unknown parameters of a theoretical probability distribution based on random variable observations. There are two main types of estimators: point estimators that provide a single value and interval estimators that provide a range of values within which the parameter is estimated to lie. Key properties for ideal estimators include being unbiased, consistent, sufficient, and having minimum variance. Examples are provided to illustrate calculating confidence intervals for population means based on sample statistics.
- The sample mean is the best estimate of the population mean and can be used to construct confidence intervals to estimate the true population mean.
- There are two situations when estimating a population mean: when the population standard deviation (σ) is known, and when σ is unknown.
- When σ is known, a z-test is used. When σ is unknown, a t-test is used since the sample standard deviation is used to estimate the population standard deviation.
This document provides an overview of key concepts related to statistical estimation and hypothesis testing, including:
- The difference between point estimation and interval estimation, and examples like confidence intervals for the mean and proportion.
- How to calculate and interpret confidence intervals.
- The roles of the null and alternative hypotheses in hypothesis testing and how to interpret p-values.
- Types I and II errors and how the significance level affects these.
- When to use parametric vs. nonparametric tests and examples of selected nonparametric tests like the chi-square test of goodness of fit.
Estimation is the process of using sample data to draw inferences about the population. A point estimate provides a single value, while an interval estimate provides a range of values expressing uncertainty. Good estimates are unbiased, meaning the expected value equals the true value, and precise, meaning the estimate is close to the true value across samples. The 95% confidence interval for a mean is calculated as the sample mean plus or minus 1.96 standard deviations, providing a 95% probability the interval contains the true mean. Similar principles apply to estimating proportions, differences between means/proportions, and small samples which use the t-distribution instead of the normal.
1. The document discusses hypothesis testing using the Z-test and T-test. It provides examples and explanations of key concepts for performing a Z-test or T-test, including defining the null and alternative hypotheses, determining critical values, calculating test statistics, and making conclusions.
2. The examples demonstrate how to perform a T-test on sample data, including calculating the sample mean and standard deviation, determining degrees of freedom, finding the critical value, computing the test statistic, and determining whether to reject the null hypothesis.
3. The document emphasizes the differences between a Z-test and T-test, notably that a Z-test is used for large samples where the population standard deviation is known, while a
Statistical Inference Part II: Types of Sampling DistributionDexlab Analytics
This is an in-depth analysis of the way different types of sampling distribution works focusing on their specific functions and interrelations as part of the discussion on the theory of sampling.
This document discusses statistical analysis and data science concepts. It covers descriptive statistics like mean, median, mode, and standard deviation. It also discusses inferential statistics including hypothesis testing, confidence intervals, and linear regression. Additionally, it discusses probability distributions, random variables, and the normal distribution. Key concepts are defined and examples are provided to illustrate statistical measures and probability calculations.
The document discusses normal and standard normal distributions. It provides examples of using a normal distribution to calculate probabilities related to bone mineral density test results. It shows how to find the probability of a z-score falling below or above certain values. It also explains how to determine the sample size needed to estimate an unknown population proportion within a given level of confidence.
The document provides an overview of descriptive statistics and statistical graphs, including measures of center such as mean, median, and mode, measures of variation such as range and standard deviation, and different types of statistical graphs like histograms, boxplots, and normal distributions. It discusses key concepts like outliers, percentiles, quartiles, sampling distributions, and the central limit theorem. The document is intended to describe important statistical tools and concepts for summarizing and describing the characteristics of data sets.
law of large number and central limit theoremlovemucheca
The document provides information about the Law of Large Numbers and the Central Limit Theorem. It discusses two key concepts:
1) As the sample size increases, the sample average converges to the population average. This is known as the Law of Large Numbers and "guarantees" stable long-term results for random events.
2) Regardless of the underlying population distribution, as sample size increases, the sample mean will be approximately normally distributed around the population mean. This is the Central Limit Theorem, which allows sample means and proportions to be analyzed using normal probability models.
The document provides examples to illustrate how these concepts can be applied, such as using the Central Limit Theorem to determine the probability that a sample average
A Lecture on Sample Size and Statistical Inference for Health ResearchersDr Arindam Basu
This document discusses concepts related to statistical inference and sample size. It begins by introducing statistical inference, estimation, and hypothesis testing. It then covers concepts of probability, including independence, mutually exclusive events, and addition. It discusses random variables and different types of variables. The document also introduces the normal distribution and central limit theorem. It provides examples of how to calculate confidence intervals and discusses interpretations of confidence intervals. Finally, it outlines the steps of hypothesis testing.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 6: Normal Probability Distribution
6.3: Sampling Distributions and Estimators
Ee184405 statistika dan stokastik statistik deskriptif 2 numerikyusufbf
Statistika adalah suatu bidang ilmu yang mempelajari cara-cara mengumpulkan data untuk selanjutnya dapat dideskripsikan dan diolah, kemudian melakukan induksi/inferensi dalam rangka membuat kesimpulan, agar dapat ditentukan keputusan yang akan diambil berdasarkan data yang dimiliki.
DATA =============> PROSES STATISTIK ===========> INFORMASI
Statistik Deskriptif adalah suatu cara menggambarkan persoalan yang berdasarkan data yang dimiliki yakni dengan cara menata data tersebut sedemikian rupa agar karakteristik data dapat dipahami dengan mudah sehingga berguna untuk keperluan selanjutnya.
According to Wikipedia point estimation involves the use of sample data to calculate a single value (known as a point estimate since it identifies a point in some parameter space) which is to serve as a "best guess" or "best estimate" of an unknown population parameter (for example, the population means).
Inferential statistics takes data from a sample and makes inferences about the larger population from which the sample was drawn.
Make use of the PPT to have a better understanding of Inferential statistics.
This document provides an overview of resampling methods, including jackknife, bootstrap, permutation, and cross-validation. It explains that resampling methods are used to approximate sampling distributions and estimate parameters' reliability when the true sampling distribution is difficult to derive. The document then describes each resampling method, their applications, and sampling procedures. It provides examples to illustrate permutation tests and how they are conducted through permutation resampling.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 7: Estimating Parameters and Determining Sample Sizes
7.3: Estimating a Population Standard Deviation or Variance
The document discusses a one-sample t-test used to compare sample data to a standard value. It provides an example comparing intelligence scores of university students to the average score of 100. The sample of 6 students had a mean of 120. Running a one-tailed t-test in SPSS, the results showed the mean score was significantly higher than 100 with t(5)=3.15, p=.02. This allows the inference that the population mean intelligence at the university is greater than the standard score of 100.
This document provides an overview of statistical estimation and inference. It discusses point estimation, which provides a single value to estimate an unknown population parameter, and interval estimation, which gives a range of plausible values for the parameter. The key aspects of interval estimation are confidence intervals, which provide a probability statement about where the true population parameter lies. The document also covers important concepts like sampling distributions, the central limit theorem, and factors that influence the width of a confidence interval like sample size. Examples are provided to demonstrate calculating point estimates, confidence intervals, and dealing with independent samples.
This article provides a brief discussion on several statistical parameters that are most commonly used in any measurement and analysis process. There are a plethora of such parameters but the most important and widely used are briefed in here.
This document discusses hypothesis testing for claims about population proportions and the difference between two population proportions. It provides information on type I and type II errors. Examples are provided to demonstrate hypothesis testing for a single proportion claim and the difference between two proportions. The examples show setting up the null and alternative hypotheses, checking assumptions, calculating the test statistic, determining the p-value or comparing to the critical value, and making a conclusion. Confidence intervals are also discussed as a way to estimate population proportions and differences between proportions. The examples provide step-by-step workings to test claims about spending behaviors with different denominations of money.
Linear regression is an approach for modeling the relationship between one dependent variable and one or more independent variables.
Algorithms to minimize the error are
OLS (Ordinary Least Square)
Gradient Descent and much more.
Let me know if anything is required. Ping me at google #bobrupakroy
Statistical estimators are functions used to estimate unknown parameters of a theoretical probability distribution based on random variable observations. There are two main types of estimators: point estimators that provide a single value and interval estimators that provide a range of values within which the parameter is estimated to lie. Key properties for ideal estimators include being unbiased, consistent, sufficient, and having minimum variance. Examples are provided to illustrate calculating confidence intervals for population means based on sample statistics.
- The sample mean is the best estimate of the population mean and can be used to construct confidence intervals to estimate the true population mean.
- There are two situations when estimating a population mean: when the population standard deviation (σ) is known, and when σ is unknown.
- When σ is known, a z-test is used. When σ is unknown, a t-test is used since the sample standard deviation is used to estimate the population standard deviation.
This document provides an overview of key concepts related to statistical estimation and hypothesis testing, including:
- The difference between point estimation and interval estimation, and examples like confidence intervals for the mean and proportion.
- How to calculate and interpret confidence intervals.
- The roles of the null and alternative hypotheses in hypothesis testing and how to interpret p-values.
- Types I and II errors and how the significance level affects these.
- When to use parametric vs. nonparametric tests and examples of selected nonparametric tests like the chi-square test of goodness of fit.
Estimation is the process of using sample data to draw inferences about the population. A point estimate provides a single value, while an interval estimate provides a range of values expressing uncertainty. Good estimates are unbiased, meaning the expected value equals the true value, and precise, meaning the estimate is close to the true value across samples. The 95% confidence interval for a mean is calculated as the sample mean plus or minus 1.96 standard deviations, providing a 95% probability the interval contains the true mean. Similar principles apply to estimating proportions, differences between means/proportions, and small samples which use the t-distribution instead of the normal.
1. The document discusses hypothesis testing using the Z-test and T-test. It provides examples and explanations of key concepts for performing a Z-test or T-test, including defining the null and alternative hypotheses, determining critical values, calculating test statistics, and making conclusions.
2. The examples demonstrate how to perform a T-test on sample data, including calculating the sample mean and standard deviation, determining degrees of freedom, finding the critical value, computing the test statistic, and determining whether to reject the null hypothesis.
3. The document emphasizes the differences between a Z-test and T-test, notably that a Z-test is used for large samples where the population standard deviation is known, while a
Statistical Inference Part II: Types of Sampling DistributionDexlab Analytics
This is an in-depth analysis of the way different types of sampling distribution works focusing on their specific functions and interrelations as part of the discussion on the theory of sampling.
This document discusses statistical analysis and data science concepts. It covers descriptive statistics like mean, median, mode, and standard deviation. It also discusses inferential statistics including hypothesis testing, confidence intervals, and linear regression. Additionally, it discusses probability distributions, random variables, and the normal distribution. Key concepts are defined and examples are provided to illustrate statistical measures and probability calculations.
The document discusses normal and standard normal distributions. It provides examples of using a normal distribution to calculate probabilities related to bone mineral density test results. It shows how to find the probability of a z-score falling below or above certain values. It also explains how to determine the sample size needed to estimate an unknown population proportion within a given level of confidence.
The document provides an overview of descriptive statistics and statistical graphs, including measures of center such as mean, median, and mode, measures of variation such as range and standard deviation, and different types of statistical graphs like histograms, boxplots, and normal distributions. It discusses key concepts like outliers, percentiles, quartiles, sampling distributions, and the central limit theorem. The document is intended to describe important statistical tools and concepts for summarizing and describing the characteristics of data sets.
law of large number and central limit theoremlovemucheca
The document provides information about the Law of Large Numbers and the Central Limit Theorem. It discusses two key concepts:
1) As the sample size increases, the sample average converges to the population average. This is known as the Law of Large Numbers and "guarantees" stable long-term results for random events.
2) Regardless of the underlying population distribution, as sample size increases, the sample mean will be approximately normally distributed around the population mean. This is the Central Limit Theorem, which allows sample means and proportions to be analyzed using normal probability models.
The document provides examples to illustrate how these concepts can be applied, such as using the Central Limit Theorem to determine the probability that a sample average
A Lecture on Sample Size and Statistical Inference for Health ResearchersDr Arindam Basu
This document discusses concepts related to statistical inference and sample size. It begins by introducing statistical inference, estimation, and hypothesis testing. It then covers concepts of probability, including independence, mutually exclusive events, and addition. It discusses random variables and different types of variables. The document also introduces the normal distribution and central limit theorem. It provides examples of how to calculate confidence intervals and discusses interpretations of confidence intervals. Finally, it outlines the steps of hypothesis testing.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 6: Normal Probability Distribution
6.3: Sampling Distributions and Estimators
Ee184405 statistika dan stokastik statistik deskriptif 2 numerikyusufbf
Statistika adalah suatu bidang ilmu yang mempelajari cara-cara mengumpulkan data untuk selanjutnya dapat dideskripsikan dan diolah, kemudian melakukan induksi/inferensi dalam rangka membuat kesimpulan, agar dapat ditentukan keputusan yang akan diambil berdasarkan data yang dimiliki.
DATA =============> PROSES STATISTIK ===========> INFORMASI
Statistik Deskriptif adalah suatu cara menggambarkan persoalan yang berdasarkan data yang dimiliki yakni dengan cara menata data tersebut sedemikian rupa agar karakteristik data dapat dipahami dengan mudah sehingga berguna untuk keperluan selanjutnya.
According to Wikipedia point estimation involves the use of sample data to calculate a single value (known as a point estimate since it identifies a point in some parameter space) which is to serve as a "best guess" or "best estimate" of an unknown population parameter (for example, the population means).
Inferential statistics takes data from a sample and makes inferences about the larger population from which the sample was drawn.
Make use of the PPT to have a better understanding of Inferential statistics.
This document provides an overview of resampling methods, including jackknife, bootstrap, permutation, and cross-validation. It explains that resampling methods are used to approximate sampling distributions and estimate parameters' reliability when the true sampling distribution is difficult to derive. The document then describes each resampling method, their applications, and sampling procedures. It provides examples to illustrate permutation tests and how they are conducted through permutation resampling.
This document provides an overview of key concepts in inferential statistics, including distributions, the normal distribution, the central limit theorem, estimators and estimates, confidence intervals, the Student's t-distribution, and formulas for calculating confidence intervals. It defines key terms and concepts, provides examples to illustrate statistical distributions and properties, and outlines the general formulas used to construct confidence intervals for different sampling situations.
Ch3_Statistical Analysis and Random Error Estimation.pdfVamshi962726
Here are the steps to solve this example:
(a) Compute the sample statistics:
Mean (x̅) = (Σxi)/n = (56.13)/10 = 5.613 cm
Standard deviation (s) = √[(Σ(xi - x̅)2)/(n-1)] = 0.6266 cm
(b) The interval over which 95% of measurements should lie is:
x̅ ± t0.025,9s = 5.613 ± 2.262(0.6266) = 5.613 ± 1.417 cm
(c) The estimated true mean value at 95% probability is:
μx = x
The document discusses the sampling distribution of means. It states that as sample size increases, the distribution of sample means approaches a normal distribution according to the Central Limit Theorem. The mean of the sampling distribution equals the population mean, and the standard deviation of the sampling distribution is the population standard deviation divided by the square root of the sample size. An example is provided to demonstrate calculating the mean and variance of a sampling distribution using a hypothetical population and different sample sizes.
1) The document discusses concepts from sampling theory including statistical inference, random sampling, sampling distributions, testing hypotheses, errors in hypothesis testing, significance levels, confidence intervals, and the t-distribution.
2) Key points include that statistical inference allows drawing conclusions about a population from a sample, random sampling is selecting a subset of individuals from a population, and sampling distributions show the distribution of sample statistics like the mean.
3) The document also discusses hypothesis testing, types of errors, significance levels, confidence intervals for estimating population parameters, and using the t-distribution to test hypotheses when the population standard deviation is unknown.
Chapter one on sampling distributions.pptFekaduAman
The document discusses sampling distributions and their properties. It introduces key concepts like population parameters, sample statistics, estimators, and the central limit theorem. It explains that as sample size increases, the sampling distribution of the sample mean approaches a normal distribution with a mean equal to the population mean and standard deviation equal to the population standard deviation divided by the square root of the sample size. The sampling distribution tells us how close sample statistics are likely to be to the corresponding population parameters.
This document provides an outline and summaries of topics related to error analysis:
- It outlines topics including binomial distribution, Poisson distribution, normal distribution, confidence interval, and least squares analysis.
- The binomial distribution section provides an example of calculating the probability of getting 2 and 3 heads out of 6 coin tosses.
- The normal distribution section explains how to calculate the probability of scoring between 90-110 on an IQ test with a mean of 100 and standard deviation of 10.
- The confidence interval section provides an example of calculating the 95% confidence interval for the population mean boiling temperature based on 6 sample measurements.
This document provides an overview of statistical inference. It discusses descriptive statistics, which summarize data, and inferential statistics, which are used to generalize from samples to populations. Key concepts covered include estimation, hypothesis testing, parameters, statistics, confidence intervals, significance levels, types of errors. Examples are given of how to calculate confidence intervals for means and proportions and how to perform hypothesis tests using z-tests and t-tests. Steps for conducting hypothesis tests are outlined.
Lect 3 background mathematics for Data Mininghktripathy
The document discusses various statistical measures used to describe data, including measures of central tendency and dispersion.
It introduces the mean, median, and mode as common measures of central tendency. The mean is the average value, the median is the middle value, and the mode is the most frequent value. It also discusses weighted means.
It then discusses various measures of data dispersion, including range, variance, standard deviation, quartiles, and interquartile range. The standard deviation specifically measures how far data values typically are from the mean and is important for describing the width of a distribution.
The document discusses basic statistical descriptions of data including measures of central tendency (mean, median, mode), dispersion (range, variance, standard deviation), and position (quartiles, percentiles). It explains how to calculate and interpret these measures. It also covers estimating these values from grouped frequency data and identifying outliers. The key goals are to better understand relationships within a data set and analyze data at multiple levels of precision.
This document provides an overview of data mining techniques discussed in Chapter 3, including parametric and nonparametric models, statistical perspectives on point estimation and error measurement, Bayes' theorem, decision trees, neural networks, genetic algorithms, and similarity measures. Nonparametric techniques like neural networks, decision trees, and genetic algorithms are particularly suitable for data mining applications involving large, dynamically changing datasets.
This document discusses moments, skewness, kurtosis, and several statistical distributions including binomial, Poisson, hypergeometric, and chi-square distributions. It defines key terms such as moment ratios, central moments, theorems, skewness, kurtosis, and correlation. Properties and applications of the binomial, Poisson, and hypergeometric distributions are provided. Finally, the document discusses the chi-square test for goodness of fit and independence.
Probability theory provides a framework for quantifying and manipulating uncertainty. It allows optimal predictions given incomplete information. The document outlines key probability concepts like sample spaces, events, axioms of probability, joint/conditional probabilities, and Bayes' rule. It also covers important probability distributions like binomial, Gaussian, and multivariate Gaussian. Finally, it discusses optimization concepts for machine learning like functions, derivatives, and using derivatives to find optima like maxima and minima.
Deep Learning Theory Seminar (Chap 3, part 2)Sangwoo Mo
This document summarizes key points from a lecture on deep learning theory:
1) It discusses the Maurey sampling technique, which shows that a finite sample approximation X^ of a random variable X converges to X as the number of samples k goes to infinity.
2) It proposes extending this technique to sample finite-width neural networks by converting the weight distribution of an infinite network to a probability measure through normalization.
3) The approximation error between outputs of the infinite and finite networks is bounded using Maurey sampling, with the bound converging to zero as the number of samples increases.
This document discusses Bayesian neural networks. It begins with an introduction to Bayesian inference and variational inference. It then explains how variational inference can be used to approximate the posterior distribution in a Bayesian neural network. Several numerical methods for obtaining the posterior distribution are covered, including Metropolis-Hastings, Hamiltonian Monte Carlo, and Stochastic Gradient Langevin Dynamics. Finally, it provides an example of classifying MNIST digits with a Bayesian neural network and analyzing model uncertainties.
The document discusses metric-based few-shot learning approaches. It introduces Matching Networks, which use an attention mechanism to calculate similarity between support and query embeddings. Prototypical Networks determine class membership for a query based on distance to prototype representations of each class. Relation Networks concatenate support and query embeddings and pass them through a relation module to predict relations as classification scores. The approaches aim to learn from few examples by leveraging metric learning in an embedding space.
Main obstacles of Bayesian statistics or Bayesian machine learning is computing posterior distribution. In many contexts, computing posterior distribution is intractable. Today, there are two main stream to detour directly computing posterior distribution. One is using sampling method(ex. MCMC) and another is Variational inference. Compared to Variational inference, MCMC takes more time and vulnerable to high-dimensional parameters. However, MCMC has strength in simplicity and guarantees of convergence. I'll briefly introduce several methods people using in application.
Texture synthesis aims to produce new texture samples from an example that are similar but not repetitive. It analyzes the example using a CNN to compute gram matrices representing the texture at different layers, then synthesizes new textures by passing noise through the CNN and minimizing differences from the example's gram matrices. Style transfer extends this to merge the texture of one image onto the content of another by matching gram matrices between layers to transfer style while preserving content. It has been shown that style and content are separable in CNN representations. Style transfer can be viewed as a type of domain adaptation between content and style domains.
Towards Deep Learning Models Resistant to Adversarial Attacks.SEMINARGROOT
This document discusses approaches to training deep neural networks to be robust against adversarial examples. It frames adversarial robustness as a minimax game between the network and an attacker. It presents projected gradient descent (PGD) and the Fast Gradient Sign Method (FGSM) as ways to solve the inner maximization problem during training. Experiments show that adversarially trained models can achieve increased robustness compared to standard networks.
Node embedding techniques learn vector representations of nodes in a graph that can be used for downstream machine learning tasks like classification, clustering, and link prediction. DeepWalk uses random walks to generate sequences of nodes that are treated similarly to sentences, and learns embeddings by predicting nodes using their neighbors, like word2vec. It does not incorporate node features or labels. Node2vec extends DeepWalk by introducing a biased random walk to learn embeddings, addressing some limitations of DeepWalk while maintaining scalability.
This document discusses graph convolutional networks (GCNs), which are neural network models for graph-structured data. GCNs aim to learn functions on graphs by preserving the graph's spatial structure and enabling weight sharing. The document outlines the basic components of a GCN, including the adjacency matrix, node features, and application of deep neural network layers. It also notes some challenges with applying convolutions to graphs and discusses approaches like using the graph Fourier transform based on the Laplacian matrix.
The document discusses different methods for denoising images in the spatial and frequency domains. It introduces spatial domain denoising techniques like mean filtering, median filtering, and adaptive filtering. It then explains how spatial domain images can be transformed into the frequency domain using Fourier and wavelet transforms. This allows denoising based on frequency content, where high frequencies associated with noise can be removed. It concludes by mentioning the CVPR Denoising Workshop as a resource.
The document discusses the importance of summarization for processing large amounts of text data. Automatic summarization systems aim to produce concise summaries while retaining the most important concepts from the original text. Current state-of-the-art models are based on deep learning and can generate multi-sentence summaries, but the accuracy and coherence of the summaries is still an area of active research.
The document contains code snippets and explanations for solving three LeetCode problems: Power of Two, Valid Parentheses, and Find Minimum in Rotated Sorted Array. For Power of Two, it provides an O(log n) solution that uses modulo and division to check if a number is a power of two. For Valid Parentheses, it provides an O(n) solution that uses a string to track opening and closing parentheses. For Find Minimum, it provides both an O(n) solution that finds the minimum by checking if each number is less than the previous, and an O(log n) solution that recursively searches halves of the array to find the minimum.
This document provides an overview of time series models and concepts. It discusses stochastic processes, stationarity, the Wold decomposition, impulse response analysis, and ARMA processes. The key points are:
1) Time series models are used to identify shocks and responses over time from stochastic processes.
2) Stationarity assumptions are needed to estimate expectations and variances from time series data using the concept that these values are time-invariant.
3) The Wold decomposition represents a stationary process as the sum of a deterministic component and stochastic prediction errors/shocks.
4) Impulse response analysis examines how past shocks continue to impact the present and future through their effect over time which decays as time
Differential Geometry for Machine LearningSEMINARGROOT
References:
Differential Geometry of Curves and Surfaces, Manfredo P. Do Carmo (2016)
Differential Geometry by Claudio Arezzo
Youtube: https://youtu.be/tKnBj7B2PSg
What is a Manifold?
Youtube: https://youtu.be/CEXSSz0gZI4
Shape analysis (MIT spring 2019) by Justin Solomon
Youtube: https://youtu.be/GEljqHZb30c
Tensor Calculus
Youtube: https://youtu.be/kGXr1SF3WmA
Manifolds: A Gentle Introduction,
Hyperbolic Geometry and Poincaré Embeddings by Brian Keng
Link: http://bjlkeng.github.io/posts/manifolds/,
http://bjlkeng.github.io/posts/hyperbolic-geometry-and-poincare-embeddings/
Statistical Learning models for Manifold-Valued measurements with application to computer vision and neuroimaging by Hyunwoo J.Kim
This document summarizes generative models like VAEs and GANs. It begins with an introduction to information theory, defining key concepts like entropy and maximum likelihood estimation. It then explains generative models as estimating the joint distribution P(X,Y) compared to discriminative models estimating P(Y|X). VAEs are discussed as maximizing the evidence lower bound (ELBO) to estimate the latent variable distribution P(Z|X), allowing generation of new X values. GANs are also covered, defining their minimax game between a generator G and discriminator D, with G learning to generate samples resembling the real data distribution Pemp.
Understanding Blackbox Prediction via Influence FunctionsSEMINARGROOT
Pang Wei Koh and Percy Liang
"Understanding Black-Box prediction via influence functions" ICML 2017 Best paper
References:
https://youtu.be/0w9fLX_T6tY
https://arxiv.org/abs/1703.04730
Attention Is All You Need (NIPS 2017)
(Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, Illia Polosukhin)
paper link: https://arxiv.org/pdf/1706.03762.pdf
Reference:
https://youtu.be/mxGCEWOxfe8 (by Minsuk Heo)
https://youtu.be/5vcj8kSwBCY (Stanford CS224N: NLP with Deep Learning | Winter 2019 | Lecture 14 – Transformers and Self-Attention)
The document discusses different types of attention mechanisms used in neural machine translation and image captioning models. It describes global attention which considers all encoder hidden states when deriving context vectors, and local attention which selectively focuses on a small window of context. Hard attention selects a single location to focus on, while soft attention takes a weighted average over locations. The document also discusses input feeding which makes the model aware of previous alignment choices.
This document is a tutorial on explainable AI from the WWW 2020 conference. It introduces explainable AI and discusses explanations from both a model and regulatory perspective. It then explores different methods for explaining individual predictions, global models, and building interpretable models. The remainder of the tutorial provides case studies on explaining diabetic retinopathy predictions, building an explainable AI engine for talent search, and using model interpretations for sales predictions. References are also included.
This document contains summaries of two LeetCode problems - Single Number and Product of Array Except Self.
For Single Number, it provides two O(n) solutions, one using a dictionary to track duplicate numbers and another using math by summing all elements and multiplying by 2, then subtracting the original sum.
For Product of Array Except Self, it again provides two O(n) solutions. The first uses a variable to track the running product and another to count zeros, updating the output array accordingly. The second avoids division by calculating left and right running products in two arrays and multiplying the values together for each output element.
This document summarizes the key steps in the locality sensitive hashing (LSH) algorithm for finding similar documents:
1. Documents are converted to sets of shingles (sequences of tokens) to represent them as high-dimensional data points.
2. MinHashing is applied to generate signatures (hashes) for each document such that similar documents are likely to have the same signatures. This compresses the data into a signature matrix.
3. LSH uses the signature matrix to hash similar documents into the same buckets with high probability, finding candidate pairs for further similarity evaluation and filtering out dissimilar pairs from consideration. This improves the computation efficiency over directly comparing all pairs.
This document discusses two algorithms for solving the Two Sum problem from LeetCode: an O(n^2) nested loop solution and an O(n) hash table solution. It also presents a coding interview question to find the maximum prime factor of a given number N and provides a solution using a while loop to iteratively check for divisibility.
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...Social Samosa
The Modern Marketing Reckoner (MMR) is a comprehensive resource packed with POVs from 60+ industry leaders on how AI is transforming the 4 key pillars of marketing – product, place, price and promotions.
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeWalaa Eldin Moustafa
Dynamic policy enforcement is becoming an increasingly important topic in today’s world where data privacy and compliance is a top priority for companies, individuals, and regulators alike. In these slides, we discuss how LinkedIn implements a powerful dynamic policy enforcement engine, called ViewShift, and integrates it within its data lake. We show the query engine architecture and how catalog implementations can automatically route table resolutions to compliance-enforcing SQL views. Such views have a set of very interesting properties: (1) They are auto-generated from declarative data annotations. (2) They respect user-level consent and preferences (3) They are context-aware, encoding a different set of transformations for different use cases (4) They are portable; while the SQL logic is only implemented in one SQL dialect, it is accessible in all engines.
#SQL #Views #Privacy #Compliance #DataLake
Build applications with generative AI on Google CloudMárton Kodok
We will explore Vertex AI - Model Garden powered experiences, we are going to learn more about the integration of these generative AI APIs. We are going to see in action what the Gemini family of generative models are for developers to build and deploy AI-driven applications. Vertex AI includes a suite of foundation models, these are referred to as the PaLM and Gemini family of generative ai models, and they come in different versions. We are going to cover how to use via API to: - execute prompts in text and chat - cover multimodal use cases with image prompts. - finetune and distill to improve knowledge domains - run function calls with foundation models to optimize them for specific tasks. At the end of the session, developers will understand how to innovate with generative AI and develop apps using the generative ai industry trends.
Codeless Generative AI Pipelines
(GenAI with Milvus)
https://ml.dssconf.pl/user.html#!/lecture/DSSML24-041a/rate
Discover the potential of real-time streaming in the context of GenAI as we delve into the intricacies of Apache NiFi and its capabilities. Learn how this tool can significantly simplify the data engineering workflow for GenAI applications, allowing you to focus on the creative aspects rather than the technical complexities. I will guide you through practical examples and use cases, showing the impact of automation on prompt building. From data ingestion to transformation and delivery, witness how Apache NiFi streamlines the entire pipeline, ensuring a smooth and hassle-free experience.
Timothy Spann
https://www.youtube.com/@FLaNK-Stack
https://medium.com/@tspann
https://www.datainmotion.dev/
milvus, unstructured data, vector database, zilliz, cloud, vectors, python, deep learning, generative ai, genai, nifi, kafka, flink, streaming, iot, edge
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
3. Which is better: machine learning or statistics?
• Prediction vs. Explanation
• Machine learning models are designed to make the most accurate
predictions possible.
• Statistical models are designed for inference about the relationships
between variables.
• Many statistical models can make predictions, but predictive accuracy is
not their strength. Likewise, machine learning models provide various
degrees of interpretability
4. 목차
1. Estimation
1) Point estimation
2) Interval estimation
3) Order Statistics
2. Hypothesis Testing
3. The Method of Monte Carlo & Bootstrap Procedure
5. Sampling and Statistics
• We have a random variable 𝑋 of interest, but its pdf 𝑓 𝑥 or pmf 𝑝(𝑥) is
not known
1. 𝑓 𝑥 or 𝑝(𝑥) is completely unknown
2. The form of 𝑓 𝑥 or 𝑝(𝑥) is know, down to a parameter 𝜃
Our information about the unknown distribution or the unknown
parameters of the distribution of 𝑋 comes from a sample on 𝑋.
1. Sampling and Statistics
6. • We consider the second classification(parametric inference)
• 𝑋 has an exponential distribution, 𝐸𝑥𝑝 𝜃 , 𝜃 is unknown.
• 𝑋 has a binomial distribution, b 𝑛, 𝑝 where 𝑛 is known, but 𝑝 is
unknown.
• 𝑋 has a gamma distribution, Γ 𝛼, 𝛽 , 𝛼 and 𝛽 are unknown.
• 𝑓 𝑥 ; 𝜃 , 𝑝 𝑥 ; 𝜃 𝜃 ∈ Ω
1. Sampling and Statistics
7. • The sample observations have the same distribution as 𝑋,
and we denote them as the random variables 𝑋1, 𝑋2, … , 𝑋 𝑛,
where 𝑛 denotes the sample size.
• When the sample is actually drawn, we use lower case letters
𝑥1, 𝑥2, … , 𝑥 𝑛 as realizations of the sample.
• Assume that 𝑋1, 𝑋2, … , 𝑋 𝑛 are mutually independent.
• Definition 1
• If 𝑋1, 𝑋2, … , 𝑋 𝑛 iid, then these random variables constitute a random
sample of size n from the common distribution.
• vs Stochastic Process ??
1. Sampling and Statistics
8. • Definition 2 (Statistic)
• 𝑋1, 𝑋2, … , 𝑋 𝑛 sample on a random variable 𝑋. Let 𝑇 = 𝑇(𝑋1, 𝑋2, … , 𝑋 𝑛) be a
function of the sample. Then 𝑇 is called a statistic(통계량)
• Once the sample is actually drawn, then 𝑡 is called the realization of 𝑇,
where 𝑡 = 𝑇(𝑥1, 𝑥2, … , 𝑥 𝑛)
• Example : 𝑋 =
𝑋1+𝑋2+⋯+𝑋 𝑛
𝑛
, 𝑆2 =
1
𝑛−1
𝑋𝑖 − 𝑋 2 are statistics
• It makes sense to consider a statistic 𝑇, which is an estimator of 𝜃.
• While we call 𝑇 an estimator of 𝜃, we call its realization 𝑡 an estimate of 𝜃.
1. Sampling and Statistics
9. Estimation
• Point estimation
• 표본평균이 15니까, 모평균도 15일것이다!
• MLE
• Interval estimation
• 표본평균이 15니까, 모평균은 (12.3, 14.7) 구간에 있을 것이다!
• Confidence Interval
1. Sampling and Statistics
10. Several Properties of point estimators
• Unbiasedness(비편향성), Consistency(일치성), efficiency(효율성)
• Definition 3 (Unbiasedness)
• 𝑋1, 𝑋2, … , 𝑋 𝑛 : sample on a random variable 𝑋 with pdf 𝑓 𝑥 ; 𝜃 , 𝜃 ∈ Ω.
Let 𝑇 = 𝑇(𝑋1, 𝑋2, … , 𝑋 𝑛) be a statistic. We say that 𝑇 is unbiased
estimator of 𝜃 if 𝐸 𝑇 = 𝜃.
1. Sampling and Statistics
11. Example : Maximum Likelihood Estimator(MLE)
• 𝐿 𝜃 = 𝑖=1
𝑛
𝑓(𝑥𝑖, 𝜃), likelihood function of the random
sample.
• Often-used estimate is that value of 𝜃 which provides a
maximum of 𝐿(𝜃). If it is unique, this is called the maximum
likelihood estimator, and denote it as 𝜃,
i.e., 𝜃 = 𝐴𝑟𝑔𝑚𝑎𝑥 𝜃 𝐿(𝜃).
1. Sampling and Statistics
12. MLE with EXPONENTIAL DISTRIBUTION
• Random sample 𝑋1, 𝑋2, … , 𝑋 𝑛 is the Γ(1, 𝜃) density. The log of
likelihood function is
𝑙 𝜃 = 𝑙𝑜𝑔
𝑖=1
𝑛
1
𝜃
𝑒
−
𝑥 𝑖
𝜃 = −𝑛log𝜃 − 𝜃−1
𝑖=1
𝑛
𝑥𝑖
•
𝜕𝑙 𝜃
𝜕𝜃
= 0,
𝜕2 𝑙 𝜃
𝜕𝜃2 < 0 yields that 𝜃 = 𝑋 is the mle of 𝜃.
• Because 𝐸 𝑋 = 𝜃, we have that 𝐸 𝑋 = 𝜃 and, hence, 𝜃 is an
unbiased estimator of 𝜃.
1. Sampling and Statistics
13. MLE with NORMAL DISTRIBUTION
• 𝑋 ~ 𝑁 𝜇, 𝜎2
∗ 𝜽 = 𝜇, 𝜎
• 𝑓 𝑥; 𝜽 =
1
2𝜋𝜎
𝑒
−
𝑥−𝜇 2
2𝜎2
• If 𝑋1, 𝑋2, … , 𝑋 𝑛 is a random variable on 𝑋, then
𝐿 𝜽 =
𝑖=1
𝑛
𝑓 𝑥𝑖; 𝜽
𝑙 𝜽 = log 𝐿 𝜽 =
𝑖=1
𝑛
log(𝑓(𝑥𝑖; 𝜃)) = −
𝑛
2
log 2𝜋 − 𝑛 log 𝜎 −
1
2
𝑖=1
𝑛
𝑥𝑖 − 𝜇
𝜎
2
𝜕𝑙
𝜕𝜇
=
𝜕𝑙
𝜕𝜎
= 0 yields 𝜇 = 𝑋 𝑎𝑛𝑑 𝜎2 = 𝑛−1
𝑖=1
𝑛
𝑋𝑖 − 𝑋 2 .
• 𝜇 is an unbiased estimator of 𝜇
• 𝜎2 is a biased estimator of 𝜎2 , though converges to 𝜎2 as → ∞.
1. Sampling and Statistics
14. 예제1(점추정)
베스킨라빈스 파인트의 무게는 정규분포를 따른다고 가정한다고 알고
있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다.
MLE에 의한 파인트 무게의 모평균의 추정값은??
표본 평균 : 321g -> 모평균의 추정값이 된다!
만약 회사 홈페이지에 310g이라고 써져있다면?
가설검정
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
2. Confidence Intervals
15. Confidence Intervals
• We discussed estimating 𝜃 by a statistic 𝜃 = 𝜃 𝑋1, 𝑋2, … , 𝑋 𝑛 ,
where 𝑋1, 𝑋2, … , 𝑋 𝑛 is a sample from the distribution of 𝑋.
• When the sample is drawn, it is unlikely that the value of 𝜃 is
the true value of the parameter.
• Error must exist.
• Embody estimate of error in terms of a confidence interval
2. Confidence Intervals
16. Confidence Intervals
• Definition 4 (Confidence Interval)
• Let 𝑋1, 𝑋2, … , 𝑋 𝑛 be a sample on a random variable 𝑋, where 𝑋 has pdf
𝑓 𝑥: 𝜃 , 𝜃 ∈ Ω. Let 𝛼 be specified. Let 𝐿 = 𝐿 𝑋1, 𝑋2, … , 𝑋 𝑛 and 𝑈 =
𝑈 𝑋1, 𝑋2, … , 𝑋 𝑛 be two statistics.
• We say that the interval (𝐿, 𝑈) is a 1 − 𝛼 100% confidence interval(신뢰
구간) for 𝜃 if
1 − 𝛼 = 𝑃 𝜃[𝜃 ∈ 𝐿, 𝑈 ]
1 − 𝛼 is called the confidence coefficient(신뢰도) of the interval
Once the sample is drawn, the realized value of the confidence interval is
𝑙, 𝑢 = (𝐿 𝑥1, 𝑥2, … , 𝑥 𝑛 , 𝑈(𝑥1, 𝑥2, … , 𝑥 𝑛))
2. Confidence Intervals
17. 예제2(구간추정)
베스킨라빈스 파인트의 무게는 정규분포를 따른다고 알고있다.
베스킨라빈스에 10번을 가서, 무게를 재봤더니 다음과 같았다.
베스킨라빈스 파인트 무게의 평균에 대한 95% 신뢰구간은?
베스킨라빈스 파인트 무게의 모평균이 얼마인지는 모르겠지만,
표본을 보니, 95% 확률로 00~00에 포함되어 있다.
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
2. Confidence Intervals
18. Confidence Interval for 𝜇 Under Normality
• Suppose 𝑋1, 𝑋2, … , 𝑋 𝑛 are a random sample from a 𝑁 𝜇, 𝜎2 . Let 𝑋 and 𝑆2 denote the sample
mean and sample variance, respectively.
𝑋 is mle of 𝜇 &
𝑛−1
𝑛
𝑆2
is mle of 𝜎2
𝑇 =
𝑋−𝜇
𝑆
𝑛
has a 𝑡-distribution with 𝑛 − 1 degrees of freedom.
1 − 𝛼 = 𝑃(−𝑡 𝛼
2,𝑛−1 < 𝑇 < 𝑡 𝛼
2,𝑛−1)
= 𝑃(−𝑡 𝛼
2,𝑛−1 <
𝑋−𝜇
𝑆
𝑛
< 𝑡 𝛼
2,𝑛−1)
= 𝑃( 𝑋 − 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
< 𝜇 < 𝑋 + 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
)
∴ 𝐿 𝑋1, 𝑋2 , … , 𝑋 𝑛 = 𝑋 − 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
and U 𝑋1, 𝑋2 , … , 𝑋 𝑛 = 𝑋 + 𝑡 𝛼
2,𝑛−1
𝑆
𝑛
2. Confidence Intervals
19. 𝑡-interval and standard error
• Once the sample is drawn, let 𝑥 and 𝑠 denote the realized
values of the statistics 𝑋 and 𝑆. Then 1 − 𝛼 100% conficence
interval for 𝜇 is given by
( 𝑥 − 𝑡 𝛼
2,𝑛−1
𝑠
𝑛
, 𝑥 + 𝑡 𝛼
2,𝑛−1
𝑠
𝑛
)
t-interval
Standard error
2. Confidence Intervals
21. Central Limit Theorem(CLT)
• The last example depends on the normality of the sampled items
• However, this is approximately true even if the sampled items are
not drawn from a normal distribution.
• Theorem 1 (Central Limit Theorem)
• Let 𝑋1, 𝑋2, … , 𝑋 𝑛 is a random sample with mean 𝜇 and finite variance 𝜎2
.
Define a random variable 𝑍 𝑛 =
𝑋−𝜇
𝜎 𝑛
. Then 𝑍 𝑛 →
𝐷
𝑍, where 𝑍~𝑁(0,1).
• 𝑋를 사용하는 경우
• 𝑛 이 충분히 큰 경우, 𝑡-분포를 사용할까? 정규분포를 사용할까?
-> 𝑡-분포 : more conservative
2. Confidence Intervals
22. Confidence Intervals for Difference in Means
• We compare the means of 𝑋 and 𝑌, denoted by 𝜇1 and 𝜇2,
respectively. Assume that the variances of 𝑋 and Y are finite
and denote them as 𝜎1
2
and 𝜎2
2
, respectively.
• Goal : Estimate ∆= 𝜇1 − 𝜇2.
• What is confidence intervals of ∆= 𝜇1 − 𝜇2?
• Sampling : 𝑋1, 𝑋2, … , 𝑋 𝑛1
& 𝑌1, 𝑌2, … , 𝑌𝑛2
• Let ∆ = 𝑋 − 𝑌. The statistic ∆ is unbiased estimator of ∆.
• 𝑉𝑎𝑟 ∆ =
𝜎1
2
𝑛1
+
𝜎2
2
𝑛2
. By CLT, 𝑍 =
∆−∆
𝑆1
2
𝑛1
+
𝑆2
2
𝑛2
~𝑁(0,12)
( 𝑥 − 𝑦 − 𝑧 𝛼 2
𝑠1
2
𝑛1
+
𝑠2
2
𝑛2
, 𝑥 − 𝑦 + 𝑧 𝛼 2
𝑠1
2
𝑛1
+
𝑠2
2
𝑛2
)
2. Confidence Intervals
23. Confidence Intervals for Difference in Means
• 𝑋 and 𝑌 are normal with the same variance; i.e., 𝜎2
= 𝜎1
2
= 𝜎2
2
.
• Estimator 𝑆 𝑝
2 of 𝜎2 is weighted average of 𝑆1
2
and 𝑆2
2
.
→ pooled estimator
→ 𝑆 𝑝
2 =
𝑛1−1 𝑆1
2+ 𝑛2−1 𝑆2
2
𝑛1+𝑛2−2
• 𝑇 =
𝑋− 𝑌 − 𝜇1−𝜇2 𝜎 𝑛1
−1+𝑛2
−1
𝑛−2 𝑆 𝑝
2 𝑛−2 𝜎2
has a 𝑡-distribution with 𝑛 − 2 degrees of freedom.
• 1 − 𝛼 100% confidence interval for ∆= 𝜇1 − 𝜇2:
• (( 𝑥- 𝑦) − 𝑡 𝛼 2,𝑛−2 𝑠 𝑝
1
𝑛1
+
1
𝑛2
, 𝑥− 𝑦 + 𝑡 𝛼 2,𝑛−2 𝑠 𝑝
1
𝑛1
+
1
𝑛2
))
2. Confidence Intervals
24. Order Statistics(순서통계량)
• 전구가 100개 있고, 𝑖 번째 전구의 수명을 확률변수 𝑋𝑖라고 하자.
첫 번째 꺼진 전구의 수명을 𝑌1, 두 번째 꺼진 전구의 수명을 𝑌2,
… , 100번째 꺼진 전구의 수명을 𝑌100 이라고 하자.
• 이렇게 순서대로 얻어진 확률변수를 순서통계량이라고 한다.
• 𝑋1, 𝑋2, … , 𝑋 𝑛이 독립이어도, 𝑌1, 𝑌2, … , 𝑌𝑛은 독립이 아닐 수 있다.
3. Order Statistics
25. Order Statistics(순서통계량)
• Theorem 2
• 𝑋 : continuous random variable whose pdf 𝑓(𝑥) has support 𝒮 =
𝑎, 𝑏 , where −∞ ≤ 𝑎 < 𝑏 ≤ ∞. Let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 denote the 𝑛
order statistics 𝑋1, 𝑋2, … , 𝑋 𝑛 from 𝑋. Then joint pdf of 𝑌1, 𝑌2, … , 𝑌𝑛 is
given by
𝑔 𝑦1, 𝑦2, … , 𝑦𝑛 =
𝑛! 𝑓 𝑦1 𝑓 𝑦2 ⋯ 𝑓 𝑦𝑛 𝑎 < 𝑦1 < 𝑦2 < ⋯ < 𝑦𝑛 < 𝑏
0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
3. Order Statistics
27. Quantiles
• Let 𝑋 be a random variable with a continuous cdf 𝐹(𝑥). For
0 < 𝑝 < 1, define the 𝒑th quantile of 𝑋 to be 𝜉 𝑝 = 𝐹−1
(𝑝).
• e.g) 𝜉0.5 is median of 𝑋.
• What is estimator of 𝜉 𝑝?
• Confidence Interval for 𝜉 𝑝?
3. Order Statistics
28. • Let 𝑋1, 𝑋2, … , 𝑋 𝑛 be a random sample from the distribution of 𝑋
and let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 be the corresponding order statistics. Let
𝑘 be the greatest integer less than or equal to 𝑝(𝑛 + 1).
• 𝐸 𝐹 𝑌𝑘 = 𝑎
𝑏
𝐹 𝑦 𝑘 𝑔 𝑘 𝑦 𝑘 𝑑𝑦 𝑘
= 0
1 𝑛!
𝑘−1 ! 𝑛−𝑘 !
𝑧 𝑘
1 − 𝑧 𝑛−𝑘
𝑑𝑧
=
𝑘
𝑛+1
≈ 𝑝
• We take 𝑌𝑘 as an estimator of the quantile 𝜉 𝑝.
• 𝑌𝑘 is called the 𝒑th sample quantile(or, 100𝒑th percentile of the sample)
• Sample quantiles are useful descriptive statistics.
3. Order Statistics
29. Application of Quantile : Five-number summary
• A five-number summary of the data consist of the following
sample quantiles
• The minimum(𝑌1), the first quartile(𝑌.25(𝑛+1) = 𝜉.25), the
median(𝑌.50(𝑛+1) = 𝜉.50), the third quartile(𝑌.75(𝑛+1) = 𝜉.75), and the
maximum(𝑌𝑛)
• Use the notation 𝑄1, 𝑄2 , and 𝑄3 to denote, respectively, the first
quartile, median, and third quartile of the sample.
3. Order Statistics
30. 예제2
• 확률변수 𝑋의 샘플 15개 데이터가 다음과 같다.
• The minimum : 𝑦1 = 56
• 𝑄1 = 𝑦4 = 94
• 𝑄2 = 𝑦8 = 102
• 𝑄3 = 𝑦12 = 108
• The maximum : 𝑦15 = 116
56 70 89 94 96 101 102 102
102 105 106 108 110 113 116
3. Order Statistics
31. Application of Quantile : Box-Whisker plot
• The five-number summary is the basis for a useful and quick plot of the
data.
• ℎ = 1.5(𝑄3 − 𝑄1)
• 𝐥𝐨𝐰𝐞𝐫 𝐟𝐞𝐧𝐜𝐞 𝐿𝐹 = 𝑄1 − ℎ, 𝐮𝐩𝐩𝐞𝐫 𝐟𝐞𝐧𝐜𝐞(𝑈𝐹) = 𝑄3 + ℎ
• Points that lie outside the fences are called potential outliers and they
are denoted by ‘o’ on the boxplot.
• The whiskers protrude from the sides of the box to what are called
adjacent points, which are the points within the fences but closest to
the fences.
3. Order Statistics
33. Application of Quantile : Diagnostic plot
• In practice, we often assume that the data follow a certain
distribution. Such an assumption needs to be checked and
there are many statistical tests which do so
Goodness of fit(위키피디아)
• We discuss one such diagnostic plot in this regard.
• q-q plot
3. Order Statistics
34. q-q plot
• Suppose 𝑋 is a random variable with cdf 𝐹 𝑥 − 𝑎 𝑏 , where 𝐹(𝑥)
is known but 𝑎 and 𝑏 > 0 may not be.
• Let 𝑍 = (𝑋 − 𝑎) 𝑏. Then, 𝑍 has cdf 𝐹(𝑧). Let 0 < 𝑝 < 1 and let 𝜉 𝑋,𝑝
be the 𝑝th quantile of 𝑋 and 𝜉 𝑍,𝑝 be the 𝑝th quantile of 𝑍.
• 𝑝 = 𝑃 𝑋 ≤ 𝜉 𝑋,𝑝 = 𝑃[𝑍 ≤
𝜉 𝑋,𝑝−𝑎
𝑏
]
⇒ 𝜉 𝑋,𝑝 = 𝑏𝜉 𝑍,𝑝 + 𝑎
• Thus, the quantiles of 𝑋 are linearly related to the quantiles of 𝑍.
3. Order Statistics
35. q-q plot
• 𝑋1, 𝑋2, … , 𝑋 𝑛 : Random sample from 𝑋 ⇒ 𝑌1 < ⋯ < 𝑌𝑛 the order statistics.
• For 𝑘 = 1, … , 𝑛, let 𝑝 𝑘 =
𝑘
𝑛+1
. ⇒ 𝑌𝑘 is an estimator of 𝜉 𝑋,𝑝.
• Denote the corresponding quantiles of the cdf 𝐹 𝑧 by 𝜉 𝑍,𝑝 = 𝐹−1(𝑝 𝑘)
• The plot of 𝑌𝑘 versus 𝜉 𝑍,𝑝 is called a q-q plot.
3. Order Statistics
36. • Figure 4.1 contain q-q plots of three different
Distributions
• Panel C appears to be most linear. We may assume that the data were generated from a
Laplace distribution
3. Order Statistics
37. Confidence Intervals for Quantiles
• For a sample size 𝑛 on 𝑋, let 𝑌1 < 𝑌2 < ⋯ < 𝑌𝑛 be the order statistics. Let 𝑘 =
[ 𝑛 + 1 𝑝]. Then the 100𝑝th sample percentile 𝑌𝑘 is a point estimator of 𝜉 𝑝.
• We now derive distribution free confidence interval for 𝜉 𝑝.
• Let 𝑖 < 𝑛 + 1 𝑝 < 𝑗, and consider the order statistics 𝑌𝑖 < 𝑌𝑗 and the event
𝑌𝑖 < 𝜉 𝑝 ∩ 𝑌𝑗 > 𝜉 𝑝 = [적어도 i개는 𝜉 𝑝보다 작고 𝑗번째는 𝜉 𝑝보다 큰 경우]
• The probability of success is 𝑃 𝑋 < 𝜉 𝑝 = 𝐹 𝜉 𝑝 = 𝑝.
𝑃 𝑌𝑖 < 𝜉 𝑝 < 𝑌𝑗 =
𝑤=𝑖
𝑗−1
𝑛
𝑤
𝑝 𝑤
1 − 𝑝 𝑛−𝑤
(= 0.95 → 𝑌𝑖, 𝑌𝑗 )
3. Order Statistics
38. Example:
Confidence Interval for the Median.
• 𝜉.50 ∶ median of 𝐹(𝑥), 𝑄2 ∶ sample median − point estimator 𝑜𝑓𝜉.50
• Select 0 < 𝛼 < 1 and Take 𝑐 𝛼 2 to be the
𝛼
2
𝑡ℎ quantile of a binomial S ~ 𝑏(𝑛,
1
2
).
• 𝑃 𝑆 ≤ 𝑐 𝛼 2 = 𝑃 𝑆 ≥ 𝑛 − 𝑐 𝛼 2 = 𝛼 2.
• Thus it follows that
𝑃 𝑌𝑐 𝛼 2+1 < 𝜉.50 < 𝑌𝑛−𝑐 𝛼 2
= 1 − 𝛼
• Hence, when the sample is drawn,
(𝑦𝑐 𝛼 2+1, 𝑦𝑛−𝑐 𝛼 2
) is a (1 − 𝛼)100% confidence interval for 𝜉.50.
3. Order Statistics
39. 예제3
• 모기 4마리를 채집해서 모기 4마리의 수명을 관찰했다. 모기의
수명이 평균이 1인 지수분포를 따를 때, 제일 오래 산 모기의 수
명이 3 이상일 확률은?
• Let 𝑌1 < 𝑌2 < 𝑌3 < 𝑌4 be the order statistics of a random
sample of size 4 from the distribution having pdf 𝑓 𝑥 = 𝑒−𝑥,
0 < 𝑥 < ∞. Find 𝑃 3 ≤ 𝑌4 .
• 𝑃 𝑌4 ≤ 𝑡 = 1 − 𝑒−𝑡 4
∴ 𝑃 𝑌4 ≥ 3 = 1 − 1 − 𝑒−3 4 = 0.1848
3. Order Statistics
40. 예제4
• 모기 4마리를 채집해서 모기 4마리의 수명을 관찰했다. 모기의
수명이 0에서 1의 값을 가지는 균등분포를 따를 때, 제일 오래
산 모기와 제일 빨리 죽은 모기의 수명의 차이가 0.5 미만일 확
률은?
• Let 𝑌1 < 𝑌2 < 𝑌3 < 𝑌4 be the order statistics of a random
sample of size 4 from the distribution having pdf 𝑓 𝑥 = 1,
0 < 𝑥 < 1. Find the probability that the range of the random
sample is less then
1
2
.
3. Order Statistics
42. Introduction to Hypothesis Testing
4 Introduction to Hypothesis Testing
Statistical Inference
Estimation
Hypothesis Testing
Point estimation
Interval estimation
43. 예제5(점추정/가설검정)
베스킨라빈스 파인트의 무게는 정규분포를 따른다고 가정한다. 베스킨라빈스에 10
번을 가서, 무게를 재봤더니 다음과 같았다.
MLE에 의한 파인트 무게의 모평균의 추정값은??
표본 평균 : 321g -> 모평균의 추정값이 된다!
만약 회사 홈페이지에 310g이라고 써져있다면?
가설검정
1 2 3 4 5 6 7 8 9 10
310 315 330 320 325 325 310 315 320 340
4 Introduction to Hypothesis Testing
44. • 귀무가설 : 파인트 무게는 평균 310g이다.
• 연구가설 : 파인트 무게는 평균 310g이 아니다.
• 모평균의 추정값 : 321g -> 귀무가설 기각 (X)
• 기각역 설정 -> 추정값이 기각역에 포함??(O)
1 2 3 4 5 6 7 8 9 10 평균
310 315 330 320 325 325 310 315 320 340 321
4 Introduction to Hypothesis Testing
45. 모분포 : 평균 100, 분산 30
Sampling
표본 평균 : 90
50개
표본 평균 : 80
표본 평균 : 40
표본 평균 : 10
표본 평균 : 1
표본 평균 : 150
귀무가설
귀무가설의 모평균에서 멀어질수록
귀무가설이 틀렸다는 생각이 커진다.
WHY?
50개 뽑았으면 우연이 아니라 필연!
4 Introduction to Hypothesis Testing
46. Hypothesis
• 𝐻0 : Null hypothesis(귀무가설)
• Represents no change or no difference from the past
• 𝐻1 : Alternative hypothesis(대립가설/연구가설)
• Represents change or difference
• Research worker’s hypothesis
𝐻0 ∶ 𝜃 ∈ 𝜔0 versus 𝐻1 ∶ 𝜃 ∈ 𝜔1
(where 𝜔0 ∪ 𝜔1 = Ω , and 𝜔0 ∩ 𝜔1 = ∅)
• The decision rule to take 𝐻0 or 𝐻0 is based on a sample 𝑋1, 𝑋2, … , 𝑋 𝑛
from 𝑋
• Hence, the decision could wrong.
4 Introduction to Hypothesis Testing
47. Decision Rule
• 𝑋1, … , 𝑋 𝑛 : random sample from the distribution of 𝑋.
• Consider testing the hypotheses :
• 𝐻0 ∶ 𝜃 ∈ 𝜔0 versus 𝐻1 ∶ 𝜃 ∈ 𝜔1
• Denote the space of the sample by 𝒟 = space{(𝑋1, … , 𝑋 𝑛)}.
• Definition 5 (Statistical Test, Critical Region)
• A test of 𝐻0 versus 𝐻1 is based on a subset 𝐶 of 𝒟.
• This set 𝐶 is called the critical region(기각역) and its corresponding
decision rule is:
Reject 𝐻0 if 𝑋1, … , 𝑋 𝑛 ∈ 𝐶
Retain 𝐻0 if 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 𝑐
4 Introduction to Hypothesis Testing
48. • the decision could wrong.
• Type 1, Type 2 error
• Goal : Select a critical region from all possible critical regions which minimizes
the probabilities of these two errors.
• in general, not possible (trade-off)
• Procedure
• Select critical regions which bound the probability of Type 1 error
• Among these critical regions we try to select one which minimizes Type 2 error.
4 Introduction to Hypothesis Testing
49. • Definition 5 (Critical region of size 𝛼)
• We say a critical region 𝐶 is of size 𝛼 if
𝛼 = max
𝜃∈𝜔0
𝑃 𝜃[ 𝑋1, … 𝑋 𝑛 ∈ 𝐶]
• Over all critical region of size 𝛼, we want to consider critical
regions which have lower probabilities of Type 2 error.
• 𝑃 𝜃 Type 2 Error = 𝑃 𝜃 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 𝑐
, 𝜃 ∈ 𝜔1
• we desire to maximize 𝑃 𝜃[ 𝑋1, … , 𝑋 𝑛 ∈ 𝐶]. (power of the test at 𝜃)
• Definition 6 (Power function)
• Power function of a critical region to be
𝛾 𝐶 𝜃 = 𝑃 𝜃 𝑋1, … , 𝑋 𝑛 ∈ 𝐶 ; 𝜃 ∈ 𝜔1
4 Introduction to Hypothesis Testing
50. Better critical region
• Given two critical regions 𝐶1 and 𝐶2 which are both of size 𝛼,
𝐶1 is better than 𝐶2 if 𝛾 𝐶1
𝜃 ≥ 𝛾 𝐶2
(𝜃) for all 𝜃 ∈ 𝜔1.
• Type1 에러의 최대값이 𝛼로 같으면, Type2 에러가 작은 것이 더 좋은
기각역이다.
4 Introduction to Hypothesis Testing
51. Test for a Binomial Proportion of Success
• 𝑋~𝐵𝑒𝑟(𝑝), 𝑝0 : probability of dying with some standard treatment
• We want to test at size 𝛼,
𝐻0 ∶ 𝑝 = 𝑝0 versus 𝐻1 ∶ 𝑝 < 𝑝0
• Let 𝑋1, … , 𝑋 𝑛 be a random sample from the distribution of X and
let 𝑆 = 𝑖=1
𝑛
𝑋𝑖 .
• An intuitive decision rule is :
Reject 𝐻0 in favor of 𝐻1 if 𝑆 ≤ 𝑘,
where 𝑘 is such that 𝛼 = 𝑃 𝐻0
[𝑆 ≤ 𝑘].
• For example, suppose 𝑛 = 20, 𝑝0 = 0.7, and 𝛼=0.15
• 𝑃 𝐻0
𝑆 ≤ 11 = 0.1133 and 𝑃 𝐻0
𝑆 ≤ 12 = 0.2277 ⋯ 𝑘 = 11? 12?
• On the conservative side, choose 𝑘 to be 11 and 𝛼 = 0.1133.
• 𝛾 𝑝 = 𝑃𝑝 𝑆 ≤ 𝑘 , 𝑝 < 𝑝0
4 Introduction to Hypothesis Testing
53. Nomenclature
• 𝐻0 ∶ 𝑝 = 𝑝0 completely specifies the underlying distribution, it
is called a simple hypothesis. ( 𝜔0 = 1)
• 𝐻1 ∶ 𝑝 < 𝑝0 is composite hypothesis.
• Frequently, 𝛼 is also called the significance level(유의수준) of
the test
• Or, ”maximum of probabilities of committing an error of Type 1 error”
• Or, “maximum of the power of the test when 𝐻0 is true”
4 Introduction to Hypothesis Testing
54. 𝒕-test : Test for 𝜇 Under Normality
• Let 𝑋 have a 𝑁(𝜇, 𝜎2) distribution. Consider the hypotheses
𝐻0 ∶ 𝜇 = 𝜇0 𝑣𝑒𝑟𝑠𝑢𝑠 𝐻1 ∶ 𝜇 > 𝜇0
• Assume that the desired size of the test is 𝛼, for 0 < 𝛼 < 1.
• Our intuitive rejection rule is to reject 𝐻0 if 𝑋 is much larger than 𝜇0.
• 𝑇 =
𝑋−𝜇0
𝑆 𝑛
, 𝑡-distribution with 𝑛-1 degree of freedom. It follows that this rejection rule
has exact level 𝛼 :
Reject 𝐻0 if 𝑇 =
𝑋 − 𝜇0
𝑆 𝑛
≥ 𝑡 𝛼,𝑛−1, where 𝛼 = 𝑃[𝑇 > 𝑡 𝛼,𝑛−1]
4 Introduction to Hypothesis Testing
55. • Large sample rule ??
• In practice, we may not be willing to assume that the
population is normal. Usually 𝑡-critical values are larger than
𝑧-critical values
• Hence, the 𝑡-test is conservative relative to the large sample
test. So, in practice, many statisticians often use the 𝑡-test.
4 Introduction to Hypothesis Testing
56. 예제6(가설검정 𝑡-test)
베스킨라빈스 파인트의 무게는 모평균이 310g 인 정규분포를 따
른다고 알고있다. 베스킨라빈스에 10번을 가서, 무게를 재봤더니
다음과 같았다.
가설 설정 :
• 귀무가설(𝐻0) : 𝜇 = 310
• 연구가설(𝐻1) : 𝜇 > 310
1 2 3 4 5 6 7 8 9 10 평균
310 315 330 320 325 325 310 315 320 340 321
4 Introduction to Hypothesis Testing
57. • 기각역 설정(유의수준(𝛼) = 5%)
• Intuitive rejection rule :
• 𝑋 much larger than 310 or much smaller than 310
• 𝑇 =
𝑋−310
𝑆 10
≥ 𝑡0.05,9
• Critical region of size 0.05
• 𝐶 = { 𝑥 ≥ 312.1170 }
• 𝑥 = 321 -> 귀묵가설 기각
• Conclusion
• 베스킨라빈스 파인트 모평균은 310g이 아니고, 310g 보다 크다.
4 Introduction to Hypothesis Testing
58. Randomized tests and 𝑝-value
• Let 𝑋1, … , 𝑋10 be random sample of size 𝑛=10 from a Poisson
distribution with mean 𝜃.
• A critical region for testing 𝐻0 ∶ 𝜃 = 0.1 against 𝐻1 ∶ 𝜃 > 0.1 is
given by 𝑌 = 𝑖=1
10
𝑋𝑖 ≥ 3. The statistic 𝑌 has a Poisson distribution
with mean 10𝜃.
• Significance level of the test is
𝑃 𝑌 ≥ 3 = 1 − 𝑃 𝑌 ≤ 2 = 1 − 0.920 = 0.080
• If the critical region defined by 𝑖=1
10
𝑋𝑖 ≥ 4 is used, the significance level is
𝛼 = 𝑃 𝑌 ≥ 4 = 1 − 𝑃 𝑌 ≤ 3 = 0.019
How we achieve a significance level of 𝛼 = 0.05??
4 Introduction to Hypothesis Testing
59. • Let 𝑊 have a Bernoulli distribution with probability of success
equal to
𝑃 𝑊 = 1 =
0.050 − 0.019
0.080 − 0.019
=
31
61
• Assume that 𝑊 is selected independently of the sample. Consider
the rejection rule
Reject 𝐻0 if 1
10
𝑥𝑖 ≥ 4 or if 1
10
𝑥𝑖 = 3 and 𝑊 = 1.
• The significance level of this rule is
𝑃 𝐻0
𝑌 ≥ 4 + 𝑃 𝑌 = 3 ∩ 𝑊 = 1
= 𝑃 𝐻0
𝑌 ≥ 4 + 𝑃 𝐻0
𝑌 = 3 𝑃 𝑊 = 1
= 0.019 + 0.061
31
61
= 0.05
• The process of performing the auxiliary experiment is referred to
as a randomized test.
4 Introduction to Hypothesis Testing
60. Observed Significance Level
• Not many statisticians like randomized tests in practice,
because the use of them means that two statisticians could
make the same assumptions, observe the same data, and yet
make different decisions.
• As a matter of fact, many statisticians report what are
commonly called observed significance level or 𝒑-values.
유의수준을 정했는데 그에 따른 기각역을 정할 수 없으면 그냥 p-value를 살피자.
4 Introduction to Hypothesis Testing
61. 모분포 : 평균 100, 분산 30
Sampling
50개
표본 평균 : 30
표본 평균 : 180
귀무가설
두 표본이 얼마나 귀무가설을
벗어나는지 측정하고 싶다.
표본평균 30이 뽑히는 것만큼
극단적으로 관찰될 확률은?
표본평균 180이 뽑히는 것만큼
극단적으로 관찰될 확률은?
유의수준을 정해준 것은 귀무가설을 얼마나 벗어나면 기각을 시킬지 미리 정해준것!
4 Introduction to Hypothesis Testing
P-value
62. • 𝑝-value is the observed “tail” probability of a statistic being at
least as extreme as the particular observed value when 𝐻0 is
true.
• If 𝑌 = 𝑢(𝑋1, 𝑋2, … , 𝑋 𝑛) is the statistic to be used in a test of
𝐻0 and if the critical region is of the form
𝑢 𝑥1, 𝑥2, … , 𝑥 𝑛 ≤ 𝑐,
an observed value 𝑢 𝑥1, 𝑥2, … , 𝑥 𝑛 = 𝑑 would mean that the
𝑝-value = 𝑃(𝑌 ≤ 𝑑; 𝐻0)
4 Introduction to Hypothesis Testing
63. 예제7
• 𝑋1, 𝑋2, … 𝑋25 a random sample from 𝑁(𝜇, 𝜎2 = 4). 𝑋~𝑁(𝜇, 0.16)
• Test 𝐻0 ∶ 𝜇 = 77 against the one-sided alternative hypothesis
𝐻1 ∶ 𝜇 < 77
• 𝑥 = 76.1
• 𝑧 − 𝑠𝑐𝑜𝑟𝑒 : -2.25
• 𝑝 − 𝑣𝑎𝑙𝑢𝑒 ∶ ∅ −2.25 = 0.012
• If we were using a significance level of 𝛼 = 0.05, we would
reject 𝐻0 and accept 𝐻1 ∶ 𝜇 < 77 because 0.012 < 0.05
4 Introduction to Hypothesis Testing
Editor's Notes
Suppose “success” is dying from a certain disease and 𝑝 0 is the probability of dying with some standard treatment. A new treatment is used on several patients, and it is hoped that the probability of dying under this new treatment is less than 𝑝 0 .