The document discusses two approaches to statistical inference - the frequentist approach and the Bayesian approach. It focuses on explaining key concepts of the frequentist approach, including point estimates, sampling distributions, confidence intervals, and hypothesis testing. Specifically, it provides examples of how to calculate a 95% confidence interval and conduct hypothesis tests using confidence intervals to determine if differences between groups are statistically significant. The document is intended as an introduction to statistical inference from a frequentist perspective.
JavaScript String:
The String object lets you work with a series of characters; it wraps Javascript's string primitive data type with a number of helper methods.
As JavaScript automatically converts between string primitives and String objects, you can call any of the helper methods of the String object on a string primitive.
JavaScript Arrays:
The Array object lets you store multiple values in a single variable. It stores a fixed-size sequential collection of elements of the same type. An array is used to store a collection of data, but it is often more useful to think of an array as a collection of variables of the same type.
JavaScript String:
The String object lets you work with a series of characters; it wraps Javascript's string primitive data type with a number of helper methods.
As JavaScript automatically converts between string primitives and String objects, you can call any of the helper methods of the String object on a string primitive.
JavaScript Arrays:
The Array object lets you store multiple values in a single variable. It stores a fixed-size sequential collection of elements of the same type. An array is used to store a collection of data, but it is often more useful to think of an array as a collection of variables of the same type.
Digital Humanities: Role of Librarians and Libraries. The use of digital evidence & methods digital authoring, publishing, digital curation and preservation, digital use and reuse of scholarship.
This ppt is mainly for library professionals and digital humanities cohorts
Use of reference management tools in Reserch : Zotero and MendeleyDr Shalini Lihitkar
This presentation endevours to throw a light on how to use RMT in Research. Describes in detail about how to install and various functions of Zotero and Mendeley alongwith creating web account in both the software.
How to publish a paper in scopus indexed journals dr.husam abdulrasool hasanDr.HUSAM ABDULRASOOL
how to publish a paper in scopus indexed journals
Dr.HUSAM ABDULRASOOL HASAN
hussam2003hussam@yahoo.com
009647737877085
https://scholar.google.com/citations?user=yQOMOCMAAAAJ&hl=en
Inferential statistics are often used to compare the differences between the treatment groups. Inferential statistics use measurements from the sample of subjects in the experiment to compare the treatment groups and make generalizations about the larger population of subjects.
Digital Humanities: Role of Librarians and Libraries. The use of digital evidence & methods digital authoring, publishing, digital curation and preservation, digital use and reuse of scholarship.
This ppt is mainly for library professionals and digital humanities cohorts
Use of reference management tools in Reserch : Zotero and MendeleyDr Shalini Lihitkar
This presentation endevours to throw a light on how to use RMT in Research. Describes in detail about how to install and various functions of Zotero and Mendeley alongwith creating web account in both the software.
How to publish a paper in scopus indexed journals dr.husam abdulrasool hasanDr.HUSAM ABDULRASOOL
how to publish a paper in scopus indexed journals
Dr.HUSAM ABDULRASOOL HASAN
hussam2003hussam@yahoo.com
009647737877085
https://scholar.google.com/citations?user=yQOMOCMAAAAJ&hl=en
Inferential statistics are often used to compare the differences between the treatment groups. Inferential statistics use measurements from the sample of subjects in the experiment to compare the treatment groups and make generalizations about the larger population of subjects.
Confidence Interval ModuleOne of the key concepts of statist.docxmaxinesmith73660
Confidence Interval Module
One of the key concepts of statistics enabling statisticians to make incredibly accurate predictions is called the Central Limit Theorem. The Central Limit Theorem is defined in this way:
· For samples of a sufficiently large size, the real distribution of means is almost always approximately normal.
· The distribution of means gets closer and closer to normal as the sample size gets larger and larger, regardless of what the original variable looks like (positively or negatively skewed).
· In other words, the original variable does not have to be normally distributed.
· This is because, if we as eccentric researchers, drew an almost infinite number of random samples from a single population (such as the student body of NMSU), the means calculated from the many samples of that population will be normally distributed and the mean calculated from all of those samples would be a very close approximation to the true population mean. It is this very characteristic that makes it possible for us, using sound probability based sampling techniques, to make highly accurate statements about characteristics of a population based upon the statistics calculated on a sample drawn from that population.
· Furthermore, we can calculate a statistic known as the standard error of the mean (abbreviated s.e.) that describes the variability of the distribution of all possible sample means in the same way that we used the standard deviation to describe the variability of a single sample. We will use the standard error of the mean (s.e.) to calculate the statistic that is the topic of this module, the confidence interval.
The formula that we use to calculate the standard error of the mean is:
s.e. = s / √N – 1
where s = the standard deviation calculated from the sample; and
N = the sample size.
So the formula tells us that the standard error of the mean is equal to the
standard deviation divided by the square root of the sample size minus 1.
This is the preferred formula for practicing professionals as it accounts for errors that may be a function of the particular sample we have selected.
THE CONFIDENCE INTERVAL (CI)
The formula for the CI is a function of the sample size (N).
For samples sizes ≥ 100, the formula for the CI is:
CI = (the sample mean) + & - Z(s.e.).
Let’s look at an example to see how this formula works.
* Please use a pdf doc. “how to solve the problem”, I have provided for you under the “notes” link.
Example 1
Suppose that we conducted interviews with 140 randomly selected individuals (N = 140) in a large metropolitan area. We assured these individuals that their answers would remain confidential, and we asked them about their law-breaking behavior. Among other questions the individuals were asked to self-report the number of times per month they exceeded the speed limit. One of the objectives of the study was to estimate (make an inference about) the average nu.
Standard Error & Confidence Intervals.pptxhanyiasimple
Certainly! Let's delve into the concept of **standard error**.
## What Is Standard Error?
The **standard error (SE)** is a statistical measure that quantifies the **variability** between a sample statistic (such as the mean) and the corresponding population parameter. Specifically, it estimates how much the sample mean would **vary** if we were to repeat the study using **new samples** from the same population. Here are the key points:
1. **Purpose**: Standard error helps us understand how well our **sample data** represents the entire population. Even with **probability sampling**, where elements are randomly selected, some **sampling error** remains. Calculating the standard error allows us to estimate the representativeness of our sample and draw valid conclusions.
2. **High vs. Low Standard Error**:
- **High Standard Error**: Indicates that sample means are **widely spread** around the population mean. In other words, the sample may not closely represent the population.
- **Low Standard Error**: Suggests that sample means are **closely distributed** around the population mean, indicating that the sample is representative of the population.
3. **Decreasing Standard Error**:
- To decrease the standard error, **increase the sample size**. Using a large, random sample minimizes **sampling bias** and provides a more accurate estimate of the population parameter.
## Standard Error vs. Standard Deviation
- **Standard Deviation (SD)**: Describes variability **within a single sample**. It can be calculated directly from sample data.
- **Standard Error (SE)**: Estimates variability across **multiple samples** from the same population. It is an **inferential statistic** that can only be estimated (unless the true population parameter is known).
### Example:
Suppose we have a random sample of 200 students, and we calculate the mean math SAT score to be 550. In this case:
- **Sample**: The 200 students
- **Population**: All test takers in the region
The standard error helps us understand how well this sample represents the entire population's math SAT scores.
Remember, the standard error is crucial for making valid statistical inferences. By understanding it, researchers can confidently draw conclusions based on sample data. 📊🔍
If you need further clarification or have additional questions, feel free to ask! 😊
---
I've provided a concise explanation of standard error, emphasizing its importance in statistical analysis. If you'd like more details or specific examples, feel free to ask! ¹²³⁴
Source: Conversation with Copilot, 5/31/2024
(1) What Is Standard Error? | How to Calculate (Guide with Examples) - Scribbr. https://www.scribbr.com/statistics/standard-error/.
(2) Standard Error (SE) Definition: Standard Deviation in ... - Investopedia. https://www.investopedia.com/terms/s/standard-error.asp.
(3) Standard error Definition & Meaning - Merriam-Webster. https://www.merriam-webster.com/dictionary/standard%20error.
(4) Standard err
Module 7 Interval estimatorsMaster for Business Statistics.docxgilpinleeanna
Module 7
Interval estimators
Master for Business Statistics
Dane McGuckian
Topics
7.1 Interval Estimate of the Population Mean with a Known Population Standard Deviation
7.2 Sample Size Requirements for Estimating the Population Mean
7.3 Interval Estimate of the Population Mean with an Unknown Population Standard Deviation
7.4 Interval Estimate of the Population Proportion
7.5 Sample Size Requirements for Estimating the Population Proportion
7.1
Interval Estimate of the Population Mean with a Known Population Standard Deviation
Interval Estimators
Quantities like the sample mean and the sample standard deviation are called point estimators because they are single values derived from sample data that are used to estimate the value of an unknown population parameter.
The point estimators used in Statistics have some very desirable traits; however, they do not come with a measure of certainty.
In other words, there is no way to determine how close the population parameter is to a value of our point estimate. For this reason, the interval estimator was developed.
An interval estimator is a range of values derived from sample data that has a certain probability of containing the population parameter.
This probability is usually referred to as confidence, and it is the main advantage that interval estimators have over point estimators.
The confidence level for a confidence interval tells us the likelihood that a given interval will contain the target parameter we are trying to estimate.
The Meaning of “Confidence Level”
Interval estimates come with a level of confidence.
The level of confidence is specified by its confidence coefficient – it is the probability (relative frequency) that an interval estimator will enclose the target parameter when the estimator is used repeatedly a very large number of times.
The most common confidence levels are 99%, 98%, 95%, and 90%.
Example: A manufacturer takes a random sample of 40 computer chips from its production line to construct a 95% confidence interval to estimate the true average lifetime of the chip. If the manufacturer formed confidence intervals for every possible sample of 40 chips, 95% of those intervals would contain the population average.
The Meaning of “Confidence Level”
In the previous example, it is important to note that once the manufacturer has constructed a 95% confidence interval, it is no longer acceptable to state that there is a 95% chance that the interval contains the true average lifetime of the computer chip.
Prior to constructing the interval, there was a 95% chance that the random interval limits would contain the true average, but once the process of collecting the sample and constructing the interval is complete, the resulting interval either does or does not contain the true average.
Thus there is a probability of 1 or 0 that the true average is contained within the interval, not a 0.95 probability.
The interval limits are random variables because the ...
Using Python library such as numpy, scipy and pandas to carry out supervised learning operations like Support vector machine, decision tree and K-nearest neighbor.
the TCP/IP protocol suite involves several methods that enables communication of which IP addressing is one of those pertinent subjects that must be considered if communication must be successful.
Data sciences is the topnotch in our world now as it enables us to predict the future and behaviors of people and systems alike.
Hence, this course focuses on introducing the processing involved in data sciences.
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
2. Introduction
The
Frequentist
Approach
The
Bayesian
Approach
Approaches to statistical
inference
In the case of the frequentist approach, it
is assumed that there is a population,
which can be represented with several
parameters, from which we can obtain
numerous random samples.
The Bayesian approach is based on a
consideration that data are fixed, not the
result of a repeatable sampling process,
but parameters describing data can be
described probabilistically.
3. Statistical Inference: The Frequentist
Approach
The ultimate objective of
statistical inference, if we
adopt the frequentist
approach, is to produce
probable propositions
concerning population
parameters from analysis
of a sample.
The most important
classes of propositions
are as follows;
Propositions about
point estimates.
Propositions about
confidence intervals
or set estimates.
Propositions about
the acceptance or
rejection of a
hypothesis.
4. • In all these cases, the production of propositions is based on a simple
assumption: we can estimate the probability that the result represented by
the proposition has been caused by chance. The estimation of this
probability by sound methods is one of the main topics of statistics.
• The main results of classical statistics are theoretical approximations, based
on idealized models and assumptions, to measure the effect of chance on
the statistic of interest.
• Thus, concepts such as the Central Limit Theorem, the empirical sample
distribution or the t-test are central to understanding this approach.
5. • The development of modern computers has opened an
alternative strategy for measuring chance that is based on
simulation; producing computationally intensive methods
including resampling methods (such as bootstrapping),
Markov chain Monte Carlo methods, etc.
• The most interesting characteristic of these methods is that
they allow us to treat more realistic models
6. Measuring the Variability in Estimates
• Estimates produced by descriptive statistics are not
equal to the truth but they are better as more data
become available. So, it makes sense to use them as
central elements of our propositions and to measure
its variability with respect to the sample size.
7. Point Estimates
Let us consider a dataset of accidents in Barcelona in 2013.
This dataset can be downloaded from the OpenDataBCN
website, Barcelona City Hall’s open data service.
Each register in the dataset represents an accident via a
series of features: weekday, hour, address, number of dead
and injured people, etc.
This dataset will represent our population: the set of all
reported traffic accidents in Barcelona during 2013
8. Sampling Distribution of Point Estimates
• Let us suppose that we are interested in describing the daily number
of traffic accidents in the streets of Barcelona in 2013. If we have
access to the population, the computation of this parameter is a
simple operation: the total number of accidents divided by 365.
9. But now, for illustrative purposes, let us suppose that we only have
access to a limited part of the data (the sample): the number of
accidents during somedays of 2013. Can we still give an
approximation of the population mean?
The most intuitive way to go about providing such a mean is simply
to take the sample mean. The sample mean is a point estimate of
the population mean. If we can only choose one value to estimate
the population mean, then this is our best guess.
10. The Traditional Approach
• It can be mathematically shown that given n independent
observations {xi}i=1,..,n of a population with a standard deviation 𝜎𝑥,
the standard deviation of the sample mean 𝜎𝑥
−, or standard error, can
be approximated by this formula:
• 𝑆𝐸 = 𝜎𝑥/√𝑛
• The estimation is sufficiently good if n>30 and the population
distribution is not skewed. This allows us to estimate the standard
error of the sample mean even if we do not have access to the
population
11.
12. Confidence intervals
• A point estimate Θ, such as the sample mean, provides a single
plausible value for a parameter. However, as we have seen, a point
estimate is rarely perfect; usually there is some error in the estimate.
That is why we have suggested using the standard error as a measure
of its variability.
• Instead of that, a next logical step would be to provide a plausible
range of values for the parameter. A plausible range of values for the
sample parameter is called a confidence interval.
13. Confidence intervals
We will base
the definition
of confidence
interval on
two ideas:
Our point estimate is the most
plausible value of the
parameter, so it makes sense to
build the confidence interval
around the point estimate.
The plausibility of a range of
values can be defined from the
sampling distribution of the
estimate
14. Theorem
• Given a population with a finite mean μ and a finite non-zero
variance 𝜎2
, the sampling distribution of the mean
approaches a normal distribution with a mean of μ and a
variance of 𝜎2
/n as n, the sample size, increases.
15. In this case, and in order to
define an interval, we can
make use of a well-known
result from probability that
applies to normal distributions:
roughly 95% of the time our
estimate will be within 1.96
standard errors of the true
mean of the distribution.
If the interval spreads out 1.96
standard errors from a
normally distributed point
estimate, intuitively we can say
that we are roughly 95%
confident that we have
captured the true parameter.
CI=[Θ−1.96×SE,Θ+1.96×SE]
17. CI
• Suppose we want to consider confidence intervals where the
confidence level is somewhat higher than 95%: perhaps we would like
a confidence level of 99%. To create a 99% confidence interval,
change 1.96 in the 95% confidence interval formula to be 2.58 (it can
be shown that 99% of the time a normal random variable will be
within 2.58 standard deviations of the mean).
• In general, if the point estimate follows the normal model with
standard errorSE, then a confidence interval for the population
parameter is
• Θ±z×SE
18. Where z corresponds to the confidence level selected
Confidence
Level
90% 95% 99% 99.9%
Z value 1.65 1.96 2.58 3.291
19. How we
compute a 95%
confidence
interval
•Draw n observations with
replacement from the original
data to create a bootstrap
sample or resample.
•Calculate the mean for the
resample
Repeat the following
steps for a large
number, s, of times:
Calculate the mean of
your s value of the
sample statistic.
Calculate the standard
deviation of your s
value of the sample
statistic
Obtain the 2.5th and
97.5th percentiles of
your s values of the
sample statistic.
21. But What Does “95% Confident” Mean?
We cannot say either
that our specific
sample contains the
true parameter or that
the interval has a 95%
chance of containing
the true parameter.
That interpretation
would not be correct
under the assumptions
of traditional statistics.
In 95% of the cases,
when I compute the
95% confidence
interval from this
sample, the true
mean of the
population will fall
within the interval
defined by these
bounds: ±1.96×SE.
22. Hypothesis Testing
Let us suppose that a deeper analysis of traffic accidents in Yaounde results
in a difference between 2010 and 2013. Of course, the difference could be
caused only by chance, because of the variability of both estimates. But it
could also be the case that traffic conditions were very different in Yaounde
during the two periods and, because of that, data from the two periods can
be considered as belonging to two different populations. Then, the relevant
question is: Are the observed effects real or not?
23. Hypothesis testing
Technically, the question is usually
translated to: Were the observed
effects statistically significant?
The process of determining the
statistical significance of an effect is
called hypothesis testing.
24. This process starts by simplifying the options into two
competing hypotheses:
H0: The mean number of daily traffic accidents is the same in 2010
and 2013 (there is only one population, one true mean, and 2010
and 2013 are just different samples from the same population).
HA: The mean number of daily traffic accidents in 2010 and 2013 is
different (2010 and 2013 are two samples from two different
populations).
25. We call H0 the null hypothesis and it
represents a skeptical point of view:
the effect we have observed is due to
chance (due to the specific sample
bias). HA is the alternative hypothesis
and it represents the other point of
view: the effect is real.
The general rule of frequentist
hypothesis testing: we will not discard
H0(and hence we will not consider
HA) unless the observed effect is
implausible under H0
26. Testing Hypotheses Using Confidence
Intervals
We can use the concept
represented by confidence
intervals to measure the
plausibility of a hypothesis.
We can illustrate the
evaluation of the hypothesis
setup by comparing the mean
rate of traffic accidents in
Barcelona during 2010 and
2013:
27.
28. • This estimate suggests that in 2013 the mean rate of traffic accidents
in Barcelona was higher than it was in 2010. But is this effect
statistically significant?
• Based on our sample, the 95% confidence interval for the mean rate
of traffic accidents in Barcelona during 2013 can be calculated as
follows:
29. Because the 2010 accident rate estimate does not fall
in the range of plausible values of 2013, we say the
alternative hypothesis cannot be discarded. That is, it
cannot be ruled out that in 2013 the mean rate of
traffic accidents in Barcelona was higher than in 2010.
30. Interpreting CI Tests
Hypothesis testing is built around rejecting or failing to reject the null hypothesis. That is, we
do not rejectH0 unless we have strong evidence against it. But what precisely does strong
evidence mean? As a general rule of thumb, for those cases where the null hypothesis is
actually true, we do not want to incorrectly reject H0 more than 5% of the time. This
corresponds to a significance level of α=0.05. In this case, the correct interpretation of our test
is as follows:
• If we use a 95% confidence interval to test a problem where the null hypothesis is true, we will make an error whenever
the point estimate is at least 1.96 standard errors away from the population parameter. This happens about 5% of the
time (2.5% in each tail)
31. Testing Hypotheses Using p-Values
The p-value is the probability
of observing data at least as
favorable to the alternative
hypothesis as our current
dataset, if the null hypothesis
is true. We typically use a
summary statistic of the data
to help compute the p-value
and evaluate the hypotheses.
Usually, if P is less than 0.05
(the chance of a fluke is less
than 5%) the result is
declared statistically
significant.
It must be pointed out that
this choice is rather arbitrary
and should not be taken as a
scientific truth
32. The goal of classical hypothesis testing is to answer the question, “Given
a sample and an apparent effect, what is the probability of seeing such
an effect by chance?” Here is how we answer that question:
The first step is to quantify the size of the apparent effect by choosing a test statistic.
The second step is to define a null hypothesis, which is a model of the system based on the assumption that the apparent
effect is not realtwo periods.
The third step is to compute a p-value, which is the probability of seeing the apparent effect if the null hypothesis is true. In
our case, we would compute the difference in means, then compute the probability of seeing a difference as big, or bigger,
under the null hypothesis
The last step is to interpret the result. If the p-value is low, the effect is said to be statistically significant, which means that it is
unlikely to have occurred by chance. In this case we infer that the effect is more likely to appear in the larger population.
33.
34. P-value approximation
To approximate
the p-value ,
we can follow
the following
procedure:
Pool the distributions, generate
samples with size n and compute
the difference in the mean.
Generate samples with size n and
compute the difference in the
mean.
Count how many differences are
larger than the observed one.
37. Interpreting P-Values
• If P is less than 0.05, there are two possible conclusions: there is a real
effect or the result is an improbable fluke. Fisher’s method offers no way of
knowing which is the case.
• If the relevant question is whether accident rates are different, then it
makes sense to test the absolute difference in means. This kind of test is
called two sided because it counts both sides of the distribution of
differences.
38. Direct Approach
• The formula for the standard error of the absolute difference in two
means is similar to the formula for other standard errors. Recall that
the standard error of a single mean can be approximated by:
• 𝑆𝐸 𝑥 =𝜎/√𝑛
• The standard error of the difference of two sample means can be
constructed from the standard errors of the separate sample means:
• 𝑆𝐸 𝑥1−𝑥2
=
𝜎1
2
𝑛1
+
𝜎2
2
𝑛2
• This would allow us to define a direct test with the 95% confidence
interval
39. But Is the Effect E Real?
We do not yet have an answer for this question! We have defined a null
hypothesis H0 (the effect is not real) and we have computed the probability of
the observed effect under the null hypothesis, which is P(E|H0), where E is an
effect as big as or bigger than the apparent effect and a p-value .
We have stated that from the frequentist point of view, we cannot consider HA
unless P(E|H0) is less than an arbitrary value. But the real answer to this question
must be based on comparing P(H0|E) to P(HA|E), not on P(E|H0)! One possible
solution to these problems is to use Bayesian reasoning; an alternative to the
frequentist approach.
40. No matter how many data you have, you will still
depend on intuition to decide how to interpret,
explain, and use that data. Data cannot speak by
themselves. Data scientists are interpreters, offering
one interpretation of what the useful narrative story
derived from the data is, if there is one at all.