INTRODUCTION
Error analysisin research methodology involves
identifying, classifying, and analyzing errors to
improve the quality and validity of research
findings. It helps researchers understand the
sources and types of errors, and how they impact
results. This analysis allows for the refinement of
research design, data collection methods, and
statistical analysis, ultimately leading to more
reliable and accurate conclusions.
3.
UNDERSTANDING THE NATUREOF ERRORS:
Errors vs. Mistakes:
In research, errors are systematic deviations
from the truth, while mistakes are random slips
or oversights.
4.
TYPES OF ERRORS:
Errors can be broadly categorized into two:
Sampling Errors: Result from using a sample
to represent a larger population, and can be
reduced by increasing sample size and using
appropriate sampling techniques.
Non-sampling Errors: Occur from various
sources like measurement errors, data collection
errors, and processing errors.
5.
Measurement Errors:Stem from flaws in the
instruments or procedures used to collect data,
impacting the accuracy and reliability of
measurements.
Statistical Errors: Relate to the use of
inappropriate statistical methods or
misinterpretations of statistical results.
6.
SOURCES OF ERRORS:
Understanding the origins of errors is crucial for
their prevention and mitigation. Common
sources include
Instrument Errors: Faults in the measuring
instruments or devices.
Observer Errors: Inconsistencies or biases
introduced by the researchers or data collectors.
Subject Errors: Participant-related factors
influencing responses or behaviors.
Methodological Errors: Flaws in the research
design, data collection procedures, or analysis
techniques.
7.
2. IDENTIFYING ANDCLASSIFYING ERRORS:
Error Identification:
Researchers need to identify errors in their data,
which can involve examining outliers,
inconsistencies, and unusual patterns.
Error Classification:
Categorizing errors based on their nature,
source, and impact on research findings helps in
understanding their root causes and developing
targeted solutions.
8.
3. ANALYZING ANDINTERPRETING ERRORS:
Quantifying Errors:
Researchers can quantify errors using statistical
methods to determine their magnitude and
impact on results.
Interpreting Error Patterns:
Analyzing error patterns can reveal systematic
biases or flaws in the research process.
Drawing Conclusions:
Error analysis helps in interpreting research
findings with greater awareness of potential
limitations and biases.
9.
4. IMPROVING RESEARCHQUALITY:
Refining Research Design:
Identifying errors can lead to modifications in the
research design to minimize potential biases and
improve the validity of findings.
Improving Data Collection:
Enhancing data collection procedures, training
data collectors, and using appropriate
instruments can reduce measurement errors.
Selecting Appropriate Statistical Methods:
Choosing the right statistical techniques and
interpreting results correctly can minimize
statistical errors.
10.
By systematicallyanalyzing errors, researchers
can strengthen their methodologies, improve the
quality of their research, and draw more reliable
and valid conclusions. According to some
research papers, error analysis is a crucial step
in ensuring the integrity of research findings.
11.
INTRODUCTION TO SCIENTIFIC
METHODOLOGY
Scientific methodology refers to a systematic,
logical, and rational approach to discovering
knowledge, solving problems, and validating
facts. It provides a structured framework for
conducting research and ensures accuracy,
reliability, and objectivity.
12.
RULES OF SCIENTIFICMETHOD
The rules act as guidelines for conducting scientific
investigations:
Objectivity
Research should be free from personal bias and based on
facts.
Empirical Evidence
All conclusions must rely on observations and experiments.
Verifiability
Results should be reproducible by other researchers under
similar conditions.
13.
Logical Reasoning- Inferences must be based
on valid reasoning—either deductive or
inductive.
Predictability- The method should help
predict future occurrences of similar phenomena.
Simplicity - Prefer simple explanations over
complex ones, if both explain the phenomena
equally well.
14.
PRINCIPLES OF SCIENTIFIC
METHOD
Observation and Identification of Problem
Start with clear and precise observation to
identify the research problem.
Formulation of Hypothesis
Propose a tentative explanation based on prior
knowledge.
Experimentation
Conduct controlled experiments to test the
hypothesis.
15.
Analysis andInterpretation
Analyze collected data objectively and interpret
results logically.
Conclusion and Generalization
Derive conclusions that can be generalized beyond
the study sample.
Falsifiability
A hypothesis should be testable and refutable if
evidence contradicts it.
Replicability
Other researchers should be able to repeat the
study and get similar results.
16.
HYPOTHESIS
For aresearcher hypothesis is a formal question
that he intends to resolve. Thus a hypothesis
may be defined as a proposition or a set of
proposition set forth as an explanation for the
occurrence of some specified group of phenomena
either asserted merely as a provisional conjecture
to guide some investigation or accepted as highly
probable in the light of established facts
17.
CHARACTERISTICS OF HYPOTHESIS
Hypothesis should be clear and precise. If the
hypothesis is not clear and precise, the inferences
drawn on its basis cannot be taken as reliable.
Hypothesis should be capable of being tested.
Hypothesis should state relationship between
variables, if it happens to be a relational
hypothesis.
Hypothesis should be limited in scope and must be
specific.
Hypothesis should be stated as far as possible in
most simple terms so that the same is easily
understandable by all concerned.
18.
Hypothesis shouldbe consistent with most
known facts.
Hypothesis should be amenable to testing within
a reasonable time.
Hypothesis must explain the facts that gave rise
to the need for explanation.
19.
BASIC CONCEPTS CONCERNING
TESTINGOF HYPOTHESES
Null hypothesis and alternative hypothesis:
In the context of statistical analysis, we often talk
about null hypothesis and alternative hypothesis.
If we are to compare method A with method B
about its superiority and if we proceed on the
assumption that both methods are equally good,
then this assumption is termed as the null
hypothesis.
we may think that the method A is superior or
the method B is inferior, we are then stating what
is termed as alternative hypothesis.
20.
The nullhypothesis is generally symbolized as H0
and the alternative hypothesis as Ha.
The null hypothesis and the alternative
hypothesis are chosen before the sample is
drawn.
21.
Alternative hypothesisis usually the one which one
wishes to prove and the null hypothesis is the one
which one wishes to disprove. Thus, a null
hypothesis represents the hypothesis we are trying
to reject, and alternative hypothesis represents all
other possibilities.
If the rejection of a certain hypothesis when it is
actually true involves great risk, it is taken
as null hypothesis because then the probability of
rejecting it when it is true is a which is chosen very
small.
Null hypothesis should always be specific hypothesis
i.e., it should not state about or approximately a
certain value.
22.
LEVEL OF SIGNIFICANCE
The level of significance in hypothesis testing is the
probability of making a wrong decision by rejecting
a true null hypothesis. It is denoted by α and
commonly set at 0.05 (5%), 0.01 (1%), or 0.10 (10%).
For example, if the significance level is 0.05, it
means there is a 5% chance of rejecting the null
hypothesis when it is actually true. This value is
used as a threshold to decide whether the test result
is statistically significant. If the p-value is less than
or equal to the chosen significance level, the null
hypothesis is rejected; otherwise, it is accepted.
The level of significance is important because it
controls the risk of error and ensures the reliability
of research conclusions.
23.
DECISION RULE ORTEST OF
HYPOTHESIS
The decision rule or test of hypothesis is a procedure
used to decide whether to accept or reject the null
hypothesis based on sample data. It involves setting
a significance level (α), choosing the appropriate
statistical test, and determining the critical region.
If the calculated test statistic falls into the critical
region (or if the p-value is less than α), the null
hypothesis is rejected; otherwise, it is not rejected.
This rule helps ensure that decisions are made
systematically and based on statistical evidence
rather than personal judgment.
24.
TYPE I ANDTYPE II ERRORS
In hypothesis testing, a Type I error occurs when
the null hypothesis is true but is wrongly rejected.
It is also called a “false positive” and its probability
is equal to the significance level (α). For example,
concluding that a new drug works when it actually
does not is a Type I error.
A Type II error occurs when the null hypothesis is
false but is wrongly accepted. It is called a “false
negative” and its probability is denoted by β. For
example, concluding that a new drug does not work
when it actually does is a Type II error. Both errors
affect the accuracy of hypothesis testing and should
be minimized for reliable results.
25.
In thecontext of testing of hypotheses, there are
basically two types of errors we can make.
We may reject H0 when H0 is true and we may
accept H0 when in fact H0 is not true. The former is
known as Type I error and the latter as Type II
error.
Type I error means rejection of hypothesis which
should have been accepted and Type II error means
accepting the hypothesis which should have been
rejected.
Type I error (false positive )is denoted by α (alpha)
known as α error, also called the level of significance
of test; and Type II error (false negative) is denoted
by β (beta) known as β error.
26.
TWO-TAILED AND ONE-TAILED
TESTS
A two-tailed test rejects the null hypothesis if,
say, the sample mean is significantly higher or
lower than the hypothesized value of the mean of
the population. Such a test is appropriate when
the null hypothesis is some specified value and
the alternative hypothesis is a value not equal to
the specified value of the null hypothesis.
27.
PROCEDURE FOR HYPOTHESIS
TESTING
To test a hypothesis means to tell (on the basis of
the data the researcher has collected) whether or
not the hypothesis seems to be valid.
1. Making a formal statement
2. Selecting a significance level
3. Deciding the distribution to use
4. Selecting a random sample and computing an
appropriate value
5. Calculation of the probability
6. Comparing the probability
28.
MAKING A FORMALSTATEMENT
The step consists in making a formal statement of the
null hypothesis (H0) and also of the alternative
hypothesis (Ha). This means that hypotheses should be
clearly stated, considering the nature of the research
problem.
The formulation of hypotheses is an important step
which must be accomplished with due care in
accordance with the object and nature of the problem
under consideration. It also indicates whether we
should use a one-tailed test or a two-tailed test. If Ha
is of the type greater than (or of the type lesser than),
we use a one-tailed test, but when Ha is of the type
“whether greater or smaller” then we use a two-tailed
test.
29.
SELECTING A SIGNIFICANCELEVEL
The hypotheses are tested on a pre-determined
level of significance and as such the same should
be specified. Generally, in practice, either 5%
level or 1% level is adopted for the purpose. The
factors that affect the level of significance are: (a)
the magnitude of the difference between sample
means; (b) the size of the samples; (c) the
variability of measurements within samples; and
(d) whether the hypothesis is directional or non-
directional
30.
DECIDING THE DISTRIBUTIONTO
USE
After deciding the level of significance, the next
step in hypothesis testing is to determine the
appropriate sampling distribution. The choice
generally remains between normal distribution
and the t-distribution. The rules for selecting the
correct distribution are similar to those which we
have stated earlier in the context of estimation
31.
SELECTING A RANDOMSAMPLE AND
COMPUTING AN APPROPRIATE VALUE
Another step is to select a random sample(s) and
compute an appropriate value from the sample
data concerning the test statistic utilizing the
relevant distribution. In other words, draw a
sample to furnish empirical data.
32.
CALCULATION OF THE
PROBABILITY
One has then to calculate the probability that the
sample result would diverge as widely as it has
from expectations, if the null hypothesis were in
fact true.
33.
COMPARING THE PROBABILITY
Finally comparing the probability thus calculated
with the specified value for α , the significance level.
If the calculated probability is equal to or smaller
than the α value in case of one-tailed test (and α /2 in
case of two-tailed test), then reject the null
hypothesis (i.e., accept the alternative hypothesis),
but if the calculated probability is greater, then
accept the null hypothesis. In case we reject H0, we
run a risk of (at most the level of significance)
committing an error of Type I, but if we accept H0,
then we run some risk (the size of which cannot be
specified as long as the H0 happens to be vague
rather than specific) of committing an error of Type
II.
34.
TESTS OF HYPOTHESES
Statisticians have developed several tests of
hypotheses (also known as the tests of
significance) for the purpose of testing of
hypotheses which can be classified as:
(a) Parametric tests or standard tests of
hypotheses
(b) Non-parametric tests or distribution-free
test of hypotheses.
35.
In researchmethodology, parametric and non-
parametric tests are used to analyze data and
draw conclusions. Parametric tests rely on
specific assumptions about the population
distribution (like normality) and are generally more
powerful when those assumptions are met. Non-
parametric tests, on the other hand, are
"distribution-free" and make fewer assumptions,
making them suitable for data that doesn't meet
the requirements of parametric tests or when
dealing with ordinal or ranked data.
36.
Z-TEST ,T-TEST, CHI2 -SQUARE
TEST, F TEST
Z test
z-test is based on the normal probability
distribution and is used for judging the
significance of several statistical measures,
particularly the mean.
This is a most frequently used test in research
studies. This test is used even when binomial
distribution or t-distribution is applicable on the
presumption that such a distribution tends to
approximate normal distribution as ‘n’ becomes
larger.
37.
Z TEST
z-testis generally used for comparing the mean of a
sample to some hypothesized mean for the population
in case of large sample, or when population variance
is known.
z-test is also used for judging he significance of
difference between means of two independent samples
in case of large samples, or when population variance
is known.
z-test is also used for comparing the sample
proportion to a theoretical value of population
proportion or for judging the difference in proportions
of two independent samples when n happens to be
large.
Besides, this test may be used for judging the
significance of median, mode, coefficient of correlation
and several other measures.
38.
T - TEST
t-test is based on t-distribution and is considered
an appropriate test for judging the significance of
a sample mean or for judging the significance of
difference between the means of two samples in
case of small sample(s) when population variance
is not known.
In case two samples are related, we use paired t-
test (or what is known as difference test) for
judging the significance of the mean of difference
between the two related samples.
It can also be used for judging the significance of
the coefficients of simple and partial correlations.
39.
The relevanttest statistic, t, is calculated from
the sample data and then compared with its
probable value based on t-distribution (to be read
from the table that gives probable values of t for
different levels of significance for different degrees
of freedom) at a specified level of significance for
concerning degrees of freedom for accepting or
rejecting the null hypothesis.
t-test applies only in case of small sample(s) when
population variance is unknown.
40.
CHI-SQUARE TEST
Chi2 -test is based on chi-square distribution and
as a parametric test is used for comparing a
sample variance to a theoretical population
variance.
41.
F-TEST
F-test isbased on F-distribution and is used to
compare the variance of the two-independent
samples.
This test is also used in the context of analysis of
variance (ANOVA) for judging the significance of
more than two sample means at one and the same
time.
It is also used for judging the significance of
multiple correlation coefficients.
Test statistic, F, is calculated and compared with its
probable value (to be seen in the F-ratio tables for
different degrees of freedom for greater and smaller
variances at specified level of significance) for
accepting or rejecting the null hypothesis.