This document provides an overview of nonparametric statistical methods for analyzing ranked data. It discusses the Wilcoxon rank-sum test and sign test, which are nonparametric alternatives to the t-test that do not assume a normal distribution. The document explains how to rank data values and handle ties. It also provides examples of using the sign test to compare a sample mean to a hypothesized value and interpreting the results.
This document summarizes the solutions to three one-way ANOVA problems testing claims about population means.
The first problem analyzes readability scores of three books and finds sufficient evidence to reject the claim that the means are all the same.
The second problem examines tree weights under different treatments and fails to support the claim that all treatment means are equal.
The third problem also looks at tree weights but in a different region, and finds sufficient evidence to fail to reject the claim that all treatment means are the same.
Solution to the practice test ch 10 correlation reg ch 11 gof ch12 anovaLong Beach City College
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Elementary Statistics Practice Test 5
Module 5
Chapter 10: Correlation and Regression
Chapter 11: Goodness of Fit and Contingency Tables
Chapter 12: Analysis of Variance
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Elementary Statistics Practice Test 5
Module 5
Chapter 10: Correlation and Regression
Chapter 11: Goodness of Fit and Contingency Tables
Chapter 12: Analysis of Variance
The document provides an overview of goodness-of-fit tests for multinomial experiments and contingency tables, which are used to test if observed frequency distributions fit expected distributions. It defines multinomial experiments, goodness-of-fit tests, and contingency tables, and explains how to perform tests of independence and homogeneity using chi-square tests on contingency tables. Sample problems are provided to test claims about categories of outcomes and the independence of variables in contingency tables.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 12: Analysis of Variance
12.1: One-Way ANOVA
This document defines and explains various types of regression analysis including linear, logistic, polynomial, stepwise, ridge and lasso regression. It discusses the key differences between correlation and regression. It also covers topics such as the least squares method, R-squared/coefficient of determination, adjusted R-squared, limitations of regression analysis and applications of regression analysis.
The document provides information about goodness-of-fit tests and contingency tables. It defines a goodness-of-fit test as testing whether an observed frequency distribution fits a claimed distribution. It also provides the notation, requirements, and steps to conduct a goodness-of-fit test including: defining the null and alternative hypotheses, calculating the test statistic as a chi-square value, finding the critical value, and making a decision to reject or fail to reject the null hypothesis. Several examples demonstrate how to perform goodness-of-fit tests to determine if sample data fits a claimed distribution.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 10: Correlation and Regression
10.2: Regression
This document summarizes the solutions to three one-way ANOVA problems testing claims about population means.
The first problem analyzes readability scores of three books and finds sufficient evidence to reject the claim that the means are all the same.
The second problem examines tree weights under different treatments and fails to support the claim that all treatment means are equal.
The third problem also looks at tree weights but in a different region, and finds sufficient evidence to fail to reject the claim that all treatment means are the same.
Solution to the practice test ch 10 correlation reg ch 11 gof ch12 anovaLong Beach City College
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Elementary Statistics Practice Test 5
Module 5
Chapter 10: Correlation and Regression
Chapter 11: Goodness of Fit and Contingency Tables
Chapter 12: Analysis of Variance
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Elementary Statistics Practice Test 5
Module 5
Chapter 10: Correlation and Regression
Chapter 11: Goodness of Fit and Contingency Tables
Chapter 12: Analysis of Variance
The document provides an overview of goodness-of-fit tests for multinomial experiments and contingency tables, which are used to test if observed frequency distributions fit expected distributions. It defines multinomial experiments, goodness-of-fit tests, and contingency tables, and explains how to perform tests of independence and homogeneity using chi-square tests on contingency tables. Sample problems are provided to test claims about categories of outcomes and the independence of variables in contingency tables.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 12: Analysis of Variance
12.1: One-Way ANOVA
This document defines and explains various types of regression analysis including linear, logistic, polynomial, stepwise, ridge and lasso regression. It discusses the key differences between correlation and regression. It also covers topics such as the least squares method, R-squared/coefficient of determination, adjusted R-squared, limitations of regression analysis and applications of regression analysis.
The document provides information about goodness-of-fit tests and contingency tables. It defines a goodness-of-fit test as testing whether an observed frequency distribution fits a claimed distribution. It also provides the notation, requirements, and steps to conduct a goodness-of-fit test including: defining the null and alternative hypotheses, calculating the test statistic as a chi-square value, finding the critical value, and making a decision to reject or fail to reject the null hypothesis. Several examples demonstrate how to perform goodness-of-fit tests to determine if sample data fits a claimed distribution.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 10: Correlation and Regression
10.2: Regression
Correlation and regression are statistical techniques used to analyze relationships between variables. Correlation determines the strength and direction of a relationship, while regression describes the linear relationship to predict changes in one variable based on changes in another. There are different types of correlation including simple, multiple, and partial correlation. Regression analysis determines the regression line that best fits the data to estimate values of one variable based on the other. The correlation coefficient measures the strength of linear correlation from -1 to 1, while regression coefficients are used to predict changes in the variables.
Linear regression and correlation analysis ppt @ bec domsBabasab Patil
This document introduces linear regression and correlation analysis. It discusses calculating and interpreting the correlation coefficient and linear regression equation to determine the relationship between two variables. It covers scatter plots, the assumptions of regression analysis, and using regression to predict and describe relationships in data. Key terms introduced include the correlation coefficient, linear regression model, explained and unexplained variation, and the coefficient of determination.
This document provides an overview of two-way analysis of variance (ANOVA). It explains that two-way ANOVA involves two categorical independent variables and one continuous dependent variable. The document outlines the objectives of two-way ANOVA, which are to analyze interactions between the two factors, and evaluate the effects of each factor. It then provides examples of how to set up and perform two-way ANOVA calculations and interpretations.
This document discusses random effects models and analysis of variance (ANOVA). It introduces one-way and two-way random effects ANOVA models, distinguishing between random and fixed effects. It describes how to perform inference on variance components in random effects models, including using Satterthwaite's procedure to obtain confidence intervals for variances. Mixed effects models are also introduced, where some factors are fixed and others random.
Chapter 6 simple regression and correlationRione Drevale
There is a significant positive correlation between amount of feed intake and live weight of broilers. The correlation coefficient (r) between feed intake and live weight is 0.726, which is statistically significant with p<0.017. On average, broilers gain approximately 0.5 kg of live weight for every 1 kg of feed consumed.
This document discusses sampling distributions and the central limit theorem. It defines key terms like population, statistic, and sampling distribution. It shows examples of how sampling distributions become more normal and less variable as the sample size increases. The central limit theorem states that for large sample sizes, the sampling distribution of the sample mean will be approximately normally distributed even if the population is not. It provides properties and rules for the sampling distributions of the sample mean and sample proportion.
This document discusses correlation and linear regression. It defines correlation as the analysis of the relationship between two quantitative variables. Pearson's r is used to calculate correlation and can range from -1 to 1, with values closer to those extremes indicating a stronger linear relationship. A positive r represents a positive correlation while a negative r represents an inverse relationship. The coefficient of determination (r-squared) provides the proportion of variance shared between variables. However, correlation does not imply causation. Linear regression finds the best fitting straight line through data points to predict the value of one variable based on the other.
Chapter 16: Correlation
(enhanced by VisualBee)nunngera
Correlation is a statistical method used to measure the relationship between two variables. A relationship exists when changes in one variable are accompanied by consistent changes in the other. A correlation evaluates the direction, form, and degree of the relationship. The Pearson correlation specifically measures the direction and strength of a linear relationship between two numerical variables. Other correlational methods like Spearman and point-biserial correlations can be used for ordinal or dichotomous variable relationships.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 9: Inferences from Two Samples
9.4: Two Variances or Standard Deviations
The document discusses goodness-of-fit tests for categorical data. It introduces notation for categorical variables with multiple categories and hypotheses for goodness-of-fit tests. Expected counts are calculated based on hypothesized proportions. The chi-square statistic is used to calculate test statistics and P-values are found using the chi-square distribution. Examples demonstrate applying goodness-of-fit tests to determine if variable categories occur with equal frequency.
This document discusses various measures of correlation. It defines correlation as the relationship between two variables and introduces the correlation coefficient, which ranges from -1 to 1 and indicates the strength of the relationship. It describes different types of correlation such as positive, negative, and zero correlation. It also outlines several methods for calculating the correlation coefficient, including rank difference methods, product moment correlation, and biserial correlation.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 11: Goodness-of-Fit and Contingency Tables
11.2: Contingency Tables
1. The document discusses categorical data analysis and goodness-of-fit tests. It introduces concepts such as univariate categorical data, expected counts, the chi-square test statistic, and assumptions of the chi-square test.
2. An example analyzes faculty status data from a university using a goodness-of-fit test to determine if the proportions are equal across categories. The test fails to reject the null hypothesis that the proportions are equal.
3. Tests for homogeneity and independence in two-way tables are described. Examples calculate expected counts and perform chi-square tests to compare populations' category proportions.
1) Non-parametric tests make fewer assumptions than parametric tests about the population distribution. They do not require the assumptions of normality and equal variances.
2) Some common non-parametric tests described in the document include the Mann-Whitney U test for comparing two independent samples, the Wilcoxon Rank Sum test for comparing two independent samples, and the Wilcoxon Signed Rank test for comparing two related samples.
3) The Kruskal-Wallis H test is also described, which is the non-parametric equivalent of the one-way ANOVA and can be used to compare three or more independent samples.
Overviews non-parametric and parametric approaches to (bivariate) linear correlation. See also: http://en.wikiversity.org/wiki/Survey_research_and_design_in_psychology/Lectures/Correlation
- Simple linear regression is used to predict values of one variable (dependent variable) given known values of another variable (independent variable).
- A regression line is fitted through the data points to minimize the deviations between the observed and predicted dependent variable values. The equation of this line allows predicting dependent variable values for given independent variable values.
- The coefficient of determination (R2) indicates how much of the total variation in the dependent variable is explained by the regression line. The standard error of estimate provides a measure of how far the observed data points deviate from the regression line on average.
- Prediction intervals can be constructed around predicted dependent variable values to indicate the uncertainty in predictions for a given confidence level, based on the
This document provides information about stepwise multiple regression, including:
1) Stepwise regression selects variables for inclusion in the model based on their statistical contribution to explaining variance in the dependent variable.
2) It aims to find the most parsimonious set of predictors that effectively predict the dependent variable by adding variables one at a time.
3) Validation is necessary when using stepwise regression to ensure the model developed on the training data generalizes to new data. 75/25 cross-validation is recommended.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 6: Normal Probability Distribution
6.5: Assessing Normality
Pearson Correlation, Spearman Correlation &Linear RegressionAzmi Mohd Tamil
This document discusses correlation and linear regression. It defines correlation as a statistic that measures the strength and direction of the linear relationship between two continuous variables. Positive correlation indicates that as one variable increases, so does the other. Negative correlation means the variables are inversely related. Linear regression can be used to predict a continuous outcome variable based on a continuous predictor variable using the regression equation y=a+bx. The regression line minimizes the sum of squared differences between the data points and the line. The slope coefficient b indicates the strength of the linear prediction and can be tested for significance.
This presentation discusses the application of logistic model in sports research. One can understand the model and the procedure involved in developing it if the assumptions for this analysis is satisfied.
This document summarizes the Wilcoxon signed-rank test, a nonparametric test used to test hypotheses about the median or location of a population when the assumptions of the t-test are not met. It provides an example applying the test to data on cardiac output measurements from 15 patients. The test calculates the differences between each observation and the hypothesized median, ranks the absolute values of the differences, and sums the ranks with positive and negative signs. The smaller of the two sums is the test statistic, which is compared to a critical value to determine if the null hypothesis that the population median equals the hypothesized value can be rejected.
This document summarizes several nonparametric statistical tests that do not rely on assumptions about the population distribution, including the sign test, Wilcoxon signed-rank test, median test, Mann-Whitney test, Kolmogorov-Smirnov goodness-of-fit test, and Kruskal-Wallis one-way analysis of variance. It provides details on the theoretical background, test statistics, assumptions, and procedures for the sign test for single samples and paired data. An example application of the paired sign test to a dental study is shown.
Correlation and regression are statistical techniques used to analyze relationships between variables. Correlation determines the strength and direction of a relationship, while regression describes the linear relationship to predict changes in one variable based on changes in another. There are different types of correlation including simple, multiple, and partial correlation. Regression analysis determines the regression line that best fits the data to estimate values of one variable based on the other. The correlation coefficient measures the strength of linear correlation from -1 to 1, while regression coefficients are used to predict changes in the variables.
Linear regression and correlation analysis ppt @ bec domsBabasab Patil
This document introduces linear regression and correlation analysis. It discusses calculating and interpreting the correlation coefficient and linear regression equation to determine the relationship between two variables. It covers scatter plots, the assumptions of regression analysis, and using regression to predict and describe relationships in data. Key terms introduced include the correlation coefficient, linear regression model, explained and unexplained variation, and the coefficient of determination.
This document provides an overview of two-way analysis of variance (ANOVA). It explains that two-way ANOVA involves two categorical independent variables and one continuous dependent variable. The document outlines the objectives of two-way ANOVA, which are to analyze interactions between the two factors, and evaluate the effects of each factor. It then provides examples of how to set up and perform two-way ANOVA calculations and interpretations.
This document discusses random effects models and analysis of variance (ANOVA). It introduces one-way and two-way random effects ANOVA models, distinguishing between random and fixed effects. It describes how to perform inference on variance components in random effects models, including using Satterthwaite's procedure to obtain confidence intervals for variances. Mixed effects models are also introduced, where some factors are fixed and others random.
Chapter 6 simple regression and correlationRione Drevale
There is a significant positive correlation between amount of feed intake and live weight of broilers. The correlation coefficient (r) between feed intake and live weight is 0.726, which is statistically significant with p<0.017. On average, broilers gain approximately 0.5 kg of live weight for every 1 kg of feed consumed.
This document discusses sampling distributions and the central limit theorem. It defines key terms like population, statistic, and sampling distribution. It shows examples of how sampling distributions become more normal and less variable as the sample size increases. The central limit theorem states that for large sample sizes, the sampling distribution of the sample mean will be approximately normally distributed even if the population is not. It provides properties and rules for the sampling distributions of the sample mean and sample proportion.
This document discusses correlation and linear regression. It defines correlation as the analysis of the relationship between two quantitative variables. Pearson's r is used to calculate correlation and can range from -1 to 1, with values closer to those extremes indicating a stronger linear relationship. A positive r represents a positive correlation while a negative r represents an inverse relationship. The coefficient of determination (r-squared) provides the proportion of variance shared between variables. However, correlation does not imply causation. Linear regression finds the best fitting straight line through data points to predict the value of one variable based on the other.
Chapter 16: Correlation
(enhanced by VisualBee)nunngera
Correlation is a statistical method used to measure the relationship between two variables. A relationship exists when changes in one variable are accompanied by consistent changes in the other. A correlation evaluates the direction, form, and degree of the relationship. The Pearson correlation specifically measures the direction and strength of a linear relationship between two numerical variables. Other correlational methods like Spearman and point-biserial correlations can be used for ordinal or dichotomous variable relationships.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 9: Inferences from Two Samples
9.4: Two Variances or Standard Deviations
The document discusses goodness-of-fit tests for categorical data. It introduces notation for categorical variables with multiple categories and hypotheses for goodness-of-fit tests. Expected counts are calculated based on hypothesized proportions. The chi-square statistic is used to calculate test statistics and P-values are found using the chi-square distribution. Examples demonstrate applying goodness-of-fit tests to determine if variable categories occur with equal frequency.
This document discusses various measures of correlation. It defines correlation as the relationship between two variables and introduces the correlation coefficient, which ranges from -1 to 1 and indicates the strength of the relationship. It describes different types of correlation such as positive, negative, and zero correlation. It also outlines several methods for calculating the correlation coefficient, including rank difference methods, product moment correlation, and biserial correlation.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 11: Goodness-of-Fit and Contingency Tables
11.2: Contingency Tables
1. The document discusses categorical data analysis and goodness-of-fit tests. It introduces concepts such as univariate categorical data, expected counts, the chi-square test statistic, and assumptions of the chi-square test.
2. An example analyzes faculty status data from a university using a goodness-of-fit test to determine if the proportions are equal across categories. The test fails to reject the null hypothesis that the proportions are equal.
3. Tests for homogeneity and independence in two-way tables are described. Examples calculate expected counts and perform chi-square tests to compare populations' category proportions.
1) Non-parametric tests make fewer assumptions than parametric tests about the population distribution. They do not require the assumptions of normality and equal variances.
2) Some common non-parametric tests described in the document include the Mann-Whitney U test for comparing two independent samples, the Wilcoxon Rank Sum test for comparing two independent samples, and the Wilcoxon Signed Rank test for comparing two related samples.
3) The Kruskal-Wallis H test is also described, which is the non-parametric equivalent of the one-way ANOVA and can be used to compare three or more independent samples.
Overviews non-parametric and parametric approaches to (bivariate) linear correlation. See also: http://en.wikiversity.org/wiki/Survey_research_and_design_in_psychology/Lectures/Correlation
- Simple linear regression is used to predict values of one variable (dependent variable) given known values of another variable (independent variable).
- A regression line is fitted through the data points to minimize the deviations between the observed and predicted dependent variable values. The equation of this line allows predicting dependent variable values for given independent variable values.
- The coefficient of determination (R2) indicates how much of the total variation in the dependent variable is explained by the regression line. The standard error of estimate provides a measure of how far the observed data points deviate from the regression line on average.
- Prediction intervals can be constructed around predicted dependent variable values to indicate the uncertainty in predictions for a given confidence level, based on the
This document provides information about stepwise multiple regression, including:
1) Stepwise regression selects variables for inclusion in the model based on their statistical contribution to explaining variance in the dependent variable.
2) It aims to find the most parsimonious set of predictors that effectively predict the dependent variable by adding variables one at a time.
3) Validation is necessary when using stepwise regression to ensure the model developed on the training data generalizes to new data. 75/25 cross-validation is recommended.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 6: Normal Probability Distribution
6.5: Assessing Normality
Pearson Correlation, Spearman Correlation &Linear RegressionAzmi Mohd Tamil
This document discusses correlation and linear regression. It defines correlation as a statistic that measures the strength and direction of the linear relationship between two continuous variables. Positive correlation indicates that as one variable increases, so does the other. Negative correlation means the variables are inversely related. Linear regression can be used to predict a continuous outcome variable based on a continuous predictor variable using the regression equation y=a+bx. The regression line minimizes the sum of squared differences between the data points and the line. The slope coefficient b indicates the strength of the linear prediction and can be tested for significance.
This presentation discusses the application of logistic model in sports research. One can understand the model and the procedure involved in developing it if the assumptions for this analysis is satisfied.
This document summarizes the Wilcoxon signed-rank test, a nonparametric test used to test hypotheses about the median or location of a population when the assumptions of the t-test are not met. It provides an example applying the test to data on cardiac output measurements from 15 patients. The test calculates the differences between each observation and the hypothesized median, ranks the absolute values of the differences, and sums the ranks with positive and negative signs. The smaller of the two sums is the test statistic, which is compared to a critical value to determine if the null hypothesis that the population median equals the hypothesized value can be rejected.
This document summarizes several nonparametric statistical tests that do not rely on assumptions about the population distribution, including the sign test, Wilcoxon signed-rank test, median test, Mann-Whitney test, Kolmogorov-Smirnov goodness-of-fit test, and Kruskal-Wallis one-way analysis of variance. It provides details on the theoretical background, test statistics, assumptions, and procedures for the sign test for single samples and paired data. An example application of the paired sign test to a dental study is shown.
The document provides information about hypothesis testing and the steps involved. It discusses:
1. Formulating the null and alternative hypotheses, specifying the significance level, and choosing the appropriate test statistic.
2. Examples of hypothesis tests involving means, including the z-test for large samples when the population variance is known, and the t-test for small samples when the population variance is unknown.
3. The calculations and decisions involved in conducting hypothesis tests, such as computing the test statistic, comparing it to the critical value, and determining whether to reject or fail to reject the null hypothesis.
This document provides guidelines for carrying out statistical analyses in SPSS and R using various datasets. It discusses how to replicate analyses from 2x2 tables using individual level data, and how to perform tests such as the Kappa test, McNemar's test, chi-square tests, tests for independent proportions, Fisher's exact test, Levene's test, Wilcoxon signed-rank tests, Mann-Whitney U tests, t-tests, and Q-Q plots in both SPSS and R. Instructions are provided for reading in SPSS data files into R and accessing variable values.
This document discusses non-parametric tests and how to use them to compare groups when assumptions of parametric tests are violated. It explains that non-parametric tests like the Wilcoxon and Kruskal-Wallis tests can be used when samples are small or data is not normally distributed. The Kruskal-Wallis test allows comparison of more than two groups by ranking all data and comparing mean ranks between groups. An example compares student grades under different teaching methods using both Kruskal-Wallis and ANOVA tests.
This document discusses the process of testing hypotheses. It begins by defining hypothesis testing as a way to make decisions about population characteristics based on sample data, which involves some risk of error. The key steps are outlined as:
1) Formulating the null and alternative hypotheses, with the null hypothesis stating no difference or relationship.
2) Computing a test statistic based on the sample data and selecting a significance level, usually 5%.
3) Comparing the test statistic to critical values to either reject or fail to reject the null hypothesis.
Examples are provided to demonstrate hypothesis testing for a single mean, comparing two means, and testing a claim about population characteristics using sample data and statistics.
This document discusses strategies for designing factorial experiments with multiple factors. It explains that factorial experiments involve studying the effect of varying levels of factors on a response variable. The optimal design strategy depends on whether the circumstances are unusual or normal. For normal circumstances where there is some noise and factors influence each other, a fractional factorial or full factorial design is typically best. The document provides details on analyzing the data from factorial experiments to determine if factor effects and interactions are significant. It includes examples of calculating main effects and interactions from 2-level factorial data.
Data categories are groupings of data with common characteristics or features. They are useful for managing the data because certain data may be treated differently based on their classification. Understanding the relationship and dependency between the different categories can help direct data quality effort
Marketing Research Hypothesis Testing.pptxxababid981
This document provides an overview of parametric and non-parametric hypothesis tests. It defines parametric tests as those that assume an underlying normal distribution, and lists common parametric tests like the z-test, t-test, F-test, and ANOVA. Non-parametric tests make no distributional assumptions and common examples discussed include the Mann-Whitney U test, chi-square test, and Kruskal-Wallis test. The document provides details on assumptions and procedures for conducting each of these important statistical hypothesis tests.
The document provides information on using SPSS and PSPP statistical software to analyze data and conduct statistical tests. It includes 4 lessons:
1. How to define and input data into the software.
2. How to generate descriptive statistics like measures of central tendency and variability to describe data.
3. How to examine relationships between variables using correlation, regression, and graphs.
4. How to perform statistical inference tests for means using one-sample t-tests, independent two-sample t-tests, and paired t-tests. Examples of hypotheses testing and interpreting results are provided.
This document provides an overview of parametric and non-parametric statistical tests. Parametric tests assume the data follows a known distribution (e.g. normal) while non-parametric tests make no assumptions. Common non-parametric tests covered include chi-square, sign, Mann-Whitney U, and Spearman's rank correlation. The chi-square test is described in more detail, including how to calculate chi-square values, degrees of freedom, and testing for independence and goodness of fit.
This document discusses hypothesis testing and significance tests. It defines key terms like parameters, statistics, sampling distribution, standard error, null and alternative hypotheses, type I and type II errors. It explains how to set up a hypothesis test, including choosing a significance level and critical value. Both one-tailed and two-tailed tests are described. Finally, it provides an overview of different types of significance tests for both large and small sample sizes.
C2 st lecture 10 basic statistics and the z test handoutfatima d
This document provides an overview of basic statistics concepts including averages, measures of dispersion, hypothesis testing, and the z-test. It defines the mode, median, mean, interquartile range, standard deviation, and absolute deviation. It explains how to perform a z-test including writing the null and alternative hypotheses, looking up the critical value, calculating the test statistic, and making a decision. Two examples of z-tests are provided to demonstrate the process.
- The document analyzes a dataset relating body fat percentage to various measurements to find a predictor.
- It finds that abdominal circumference has the highest correlation (0.81) to body fat percentage. A linear regression model is fitted with abdominal circumference predicting body fat percentage.
- The model is found to be statistically significant and explains 66.21% of the variability in body fat percentage based on the abdominal circumference measurement.
Data Science - Part XII - Ridge Regression, LASSO, and Elastic NetsDerek Kane
The document discusses various regression techniques including ridge regression, lasso regression, and elastic net regression. It begins with an overview of advancements in regression analysis since the late 1800s/early 1900s enabled by increased computing power. Modern high-dimensional data often has many independent variables, requiring improved regression methods. The document then provides technical explanations and formulas for ordinary least squares regression, ridge regression, lasso regression, and their properties such as bias-variance tradeoffs. It explains how ridge and lasso regression address limitations of OLS through regularization that shrinks coefficients.
Week 5 Lecture 14 The Chi Square TestQuite often, patterns of .docxcockekeshia
Week 5 Lecture 14
The Chi Square Test
Quite often, patterns of responses or measures give us a lot of information. Patterns are generally the result of counting how many things fit into a particular category. Whenever we make a histogram, bar, or pie chart we are looking at the pattern of the data. Frequently, changes in these visual patterns will be our first clues that things have changed, and the first clue that we need to initiate a research study (Lind, Marchel, & Wathen, 2008).
One of the most useful test in examining patterns and relationships in data involving counts (how many fit into this category, how many into that, etc.) is the chi-square. It is extremely easy to calculate and has many more uses than we will cover. Examining patterns involves two uses of the Chi-square - the goodness of fit and the contingency table. Both of these uses have a common trait: they involve counts per group. In fact, the chi-square is the only statistic we will look at that we use when we have counts per multiple groups (Tanner & Youssef-Morgan, 2013). Chi Square Goodness of Fit Test
The goodness of fit test checks to see if the data distribution (counts per group) matches some pattern we are interested in. Example: Are the employees in our example company distributed equal across the grades? Or, a more reasonable expectation for a company might be are the employees distributed in a pyramid fashion – most on the bottom and few at the top?
The Chi Square test compares the actual versus a proposed distribution of counts by generating a measure for each cell or count: (actual – expected)2/actual. Summing these for all of the cells or groups provides us with the Chi Square Statistic. As with our other tests, we determine the p-value of getting a result as large or larger to determine if we reject or not reject our null hypothesis. An example will show the approach using Excel.
Regardless of the Chi Square test, the chi square related functions are found in the fx Statistics window rather than the Data Analysis where we found the t and ANOVA test functions. The most important for us are:
· CHISQ.TEST (actual range, expected range) – returns the p-value for the test
· CHISQ.INV.RT(p-value, df) – returns the actual Chi Square value for the p-value or probability value used.
· CHISQ.DIST.RT(X, df) – returns the p-value for a given value.
When we have a table of actual and expected results, using the =CHISQ.TEST(actual range, expected range) will provide us with the p-value of the calculated chi square value (but does not give us the actual calculated chi square value for the test). We can compare this value against our alpha criteria (generally 0.05) to make our decision about rejecting or not rejecting the null hypothesis.
If, after finding the p-value for our chi square test, we want to determine the calculated value of the chi square statistic, we can use the =CHISQ.INV.RT(probability, df) function, the value for probability is .
Chapter 11 Chi-Square Tests and ANOVA 359 Chapter .docxbartholomeocoombs
Chapter 11: Chi-Square Tests and ANOVA
359
Chapter 11: Chi-Square and ANOVA Tests
This chapter presents material on three more hypothesis tests. One is used to determine
significant relationship between two qualitative variables, the second is used to determine
if the sample data has a particular distribution, and the last is used to determine
significant relationships between means of 3 or more samples.
Section 11.1: Chi-Square Test for Independence
Remember, qualitative data is where you collect data on individuals that are categories or
names. Then you would count how many of the individuals had particular qualities. An
example is that there is a theory that there is a relationship between breastfeeding and
autism. To determine if there is a relationship, researchers could collect the time period
that a mother breastfed her child and if that child was diagnosed with autism. Then you
would have a table containing this information. Now you want to know if each cell is
independent of each other cell. Remember, independence says that one event does not
affect another event. Here it means that having autism is independent of being breastfed.
What you really want is to see if they are not independent. In other words, does one
affect the other? If you were to do a hypothesis test, this is your alternative hypothesis
and the null hypothesis is that they are independent. There is a hypothesis test for this
and it is called the Chi-Square Test for Independence. Technically it should be called
the Chi-Square Test for Dependence, but for historical reasons it is known as the test for
independence. Just as with previous hypothesis tests, all the steps are the same except for
the assumptions and the test statistic.
Hypothesis Test for Chi-Square Test
1. State the null and alternative hypotheses and the level of significance
Ho : the two variables are independent (this means that the one variable is not
affected by the other)
HA : the two variables are dependent (this means that the one variable is affected
by the other)
Also, state your α level here.
2. State and check the assumptions for the hypothesis test
a. A random sample is taken.
b. Expected frequencies for each cell are greater than or equal to 5 (The expected
frequencies, E, will be calculated later, and this assumption means E ≥ 5 ).
3. Find the test statistic and p-value
Finding the test statistic involves several steps. First the data is collected and
counted, and then it is organized into a table (in a table each entry is called a cell).
These values are known as the observed frequencies, which the symbol for an
observed frequency is O. Each table is made up of rows and columns. Then each
row is totaled to give a row total and each column is totaled to give a column
total.
Chapter 11: Chi-Squared Tests and ANOVA
360
The null hypothesis is that the variables are independent. Using the multiplication.
R can perform many classical hypothesis tests, including χ2 tests, t-tests, ANOVA, and regression. The document provides examples of using R functions like prop.test(), chisq.test(), t.test(), aov(), and lm() to conduct hypothesis tests for proportions, independence, means comparisons, ANOVA, and regression. It also demonstrates how to obtain additional results like confidence intervals, observed and expected values, residuals, and post hoc tests.
This document provides an outline of the programme for the 33rd Annual Meeting of the International Association of Cancer Registry (IACR) taking place in Mauritius from October 10-14, 2011. The programme includes pre-meeting courses on examining cancer trends and making predictions, and training on the CanReg5 cancer registry software. The conference programme consists of oral presentations divided into sessions covering topics like cancer in Mauritius, infections and cancer, and cancer registration challenges. There will also be poster presentations, a welcome reception, and a gala dinner as part of the social programme.
Abstract
Prevalence and incidence are measures that are used for monitoring the occurrence of a disease. Prevalence can be computed from readily available cross-sectional data but incidence is traditionally computed from longitudinal data from longitudinal studies. Longitudinal studies are characterised by financial and logistical problems where as cross-sectional studies are easy to conduct. This paper introduces a new method for estimating HIV incidence from grouped cross-sectional sero-prevalence data from settings where antiretroviral therapy is provided to those who are eligible according to recommended criteria for the administration of such drugs.
Abstract
Prevalence and incidence are measures that are used for monitoring the occurrence of a disease. Prevalence can be computed from readily available cross-sectional data but incidence is traditionally computed from longitudinal data from longitudinal studies. Longitudinal studies are characterised by financial and logistical problems where as cross-sectional studies are easy to conduct. This paper introduces a new method for estimating HIV incidence from grouped cross-sectional sero-prevalence data from settings where antiretroviral therapy is provided to those who are eligible according to recommended criteria for the administration of such drugs.
- The study analyzed data from the 2000 Malawi Demographic and Health Survey to determine factors influencing willingness to undergo voluntary HIV counseling and testing among Malawians prior to marriage.
- Willingness for premarital HIV testing was positively associated with increased age, urban residence, and a preference for confidentiality of one's HIV status. However, it was negatively associated with knowledge of HIV/AIDS, testing locations, sexually transmitted infections, and a belief that abstinence prevents HIV.
- Not all population groups had an equal likelihood of accepting voluntary HIV counseling and testing. Public health interventions on HIV testing need to be tailored to different groups.
- The study analyzed data from the 2000 Malawi Demographic and Health Survey to determine factors influencing willingness to undergo voluntary HIV counseling and testing among Malawians prior to marriage.
- Willingness for premarital HIV testing was positively associated with increased age, urban residence, and a preference for confidentiality of one's HIV status. However, it was negatively associated with knowledge of HIV/AIDS, testing locations, sexually transmitted infections, and a belief that abstinence prevents HIV.
- Not all population groups had an equal likelihood of accepting voluntary HIV counseling and testing. Public health interventions on HIV testing need to be tailored to different groups.
1. The study assessed the impact of food supplementation provided by the World Food Programme to patients enrolled in a home-based care program for chronically ill patients in Malawi, most of whom had HIV/AIDS.
2. The study compared the survival and nutritional status of patients who did not receive food supplementation before July 2003 to those who received supplementation after. It found that food supplementation did not improve patient survival or nutritional status, though it had a small non-significant effect on nutritional status.
3. Providing additional oil to some families may have improved survival slightly but did not affect nutritional status. The study concludes that food supplementation was not very effective for these patients, possibly because it was introduced too late or
- The study analyzed cancer registry data from Malawi between 1996-2005 to describe the age at cancer diagnosis.
- The median ages at diagnosis were lower for AIDS-defining cancers (most 42 years) than non-AIDS defining cancers (at least 46 years).
- Childhood and adult cancer ages followed lognormal distributions. The overall age distribution was best modeled as a finite mixture of two lognormal distributions, with means of 5.1 and 45.1 years, reflecting a bimodal distribution with peaks in childhood and mid-adulthood.
This document summarizes a study of socio-demographic characteristics associated with HIV and syphilis seroreactivity among pregnant women in Blantyre, Malawi between 2000-2004. The study found that 30% of women were HIV positive and 5% were syphilis seroreactive. In multivariate analysis, HIV infection was positively associated with higher socioeconomic status, being formerly married, and older age. Syphilis seroreactivity was positively associated with rural residence, multigravidity, and previous STI diagnosis and negatively associated with higher education levels. The study demonstrates the need for improved strategies to prevent HIV and syphilis in women in Malawi.
This study examined continued professional development opportunities for healthcare workers in Blantyre, Malawi. The study found that the most common forms of continued development were workshops/seminars and clinical handover meetings. Nearly all participants had attended a workshop in the past year. However, access to professional journals was very low, with few individuals subscribing personally and few health facilities subscribing. Most participants expressed interest in receiving free journals. The study concluded there is a need to improve healthcare workers' access to relevant professional literature and for licensing boards to consider mandatory continuing education requirements.
This document summarizes a study that aimed to identify potential partners for preventing mother-to-child transmission of HIV/AIDS in Blantyre, Malawi. The study surveyed 321 pregnant women attending antenatal clinics. It found that close relatives, spouses, the media (especially radio), and health workers are important sources of health information and potential partners in prevention efforts. While most women intended to involve their mother or sister during delivery and believed their spouse would support breastfeeding, few had ever attended clinics with their spouse or discussed HIV. The study concludes these partners should be engaged in programs aimed at preventing mother-to-child transmission of HIV.
This study explored physical trauma experiences among 217 school children in Ndirande, Malawi. Many children reported experiences with trauma: 86 had fallen from a tree, with 44 being injured; 8 had been hit by a motor vehicle, with 2 hospitalized; and 87 had witnessed a road accident. Girls were more likely than boys to fall from trees and get injured. While most children reported being taught road safety, only 41.9% knew the proper procedure for crossing the road. The study highlights the exposure of Malawian children to physical trauma and the need for improved prevention, education, and management of trauma.
This study examined the effects of rotating shift work on the sleep quality and duration of nurses in Malawi. Twenty-four nurses who worked rotating shifts (day shifts, night shifts, days off) completed questionnaires on their sleep. Their sleep was significantly shorter and of lower quality during night shifts compared to days off. Even during days off, sleep problems persisted, indicating accumulated fatigue from shift work. In contrast, nurses who only worked day shifts showed no differences in sleep between work days and days off. The rotating shift schedule negatively impacted nurses' sleep, and changes to the shift system were recommended to improve nurses' sleep and well-being.
More from College of Medicine(University of Malawi) (14)
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Dr. Vinod Kumar Kanvaria
Exploiting Artificial Intelligence for Empowering Researchers and Faculty,
International FDP on Fundamentals of Research in Social Sciences
at Integral University, Lucknow, 06.06.2024
By Dr. Vinod Kumar Kanvaria
A Strategic Approach: GenAI in EducationPeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
Executive Directors Chat Leveraging AI for Diversity, Equity, and InclusionTechSoup
Let’s explore the intersection of technology and equity in the final session of our DEI series. Discover how AI tools, like ChatGPT, can be used to support and enhance your nonprofit's DEI initiatives. Participants will gain insights into practical AI applications and get tips for leveraging technology to advance their DEI goals.
The simplified electron and muon model, Oscillating Spacetime: The Foundation...RitikBhardwaj56
Discover the Simplified Electron and Muon Model: A New Wave-Based Approach to Understanding Particles delves into a groundbreaking theory that presents electrons and muons as rotating soliton waves within oscillating spacetime. Geared towards students, researchers, and science buffs, this book breaks down complex ideas into simple explanations. It covers topics such as electron waves, temporal dynamics, and the implications of this model on particle physics. With clear illustrations and easy-to-follow explanations, readers will gain a new outlook on the universe's fundamental nature.
বাংলাদেশের অর্থনৈতিক সমীক্ষা ২০২৪ [Bangladesh Economic Review 2024 Bangla.pdf] কম্পিউটার , ট্যাব ও স্মার্ট ফোন ভার্সন সহ সম্পূর্ণ বাংলা ই-বুক বা pdf বই " সুচিপত্র ...বুকমার্ক মেনু 🔖 ও হাইপার লিংক মেনু 📝👆 যুক্ত ..
আমাদের সবার জন্য খুব খুব গুরুত্বপূর্ণ একটি বই ..বিসিএস, ব্যাংক, ইউনিভার্সিটি ভর্তি ও যে কোন প্রতিযোগিতা মূলক পরীক্ষার জন্য এর খুব ইম্পরট্যান্ট একটি বিষয় ...তাছাড়া বাংলাদেশের সাম্প্রতিক যে কোন ডাটা বা তথ্য এই বইতে পাবেন ...
তাই একজন নাগরিক হিসাবে এই তথ্য গুলো আপনার জানা প্রয়োজন ...।
বিসিএস ও ব্যাংক এর লিখিত পরীক্ষা ...+এছাড়া মাধ্যমিক ও উচ্চমাধ্যমিকের স্টুডেন্টদের জন্য অনেক কাজে আসবে ...
How to Fix the Import Error in the Odoo 17Celine George
An import error occurs when a program fails to import a module or library, disrupting its execution. In languages like Python, this issue arises when the specified module cannot be found or accessed, hindering the program's functionality. Resolving import errors is crucial for maintaining smooth software operation and uninterrupted development processes.
How to Add Chatter in the odoo 17 ERP ModuleCeline George
In Odoo, the chatter is like a chat tool that helps you work together on records. You can leave notes and track things, making it easier to talk with your team and partners. Inside chatter, all communication history, activity, and changes will be displayed.
2. ii
NONPARAMETRIC STATISTICAL TESTS
Introduction
When we use t-tests, the assumption we make about the data is that it is normally
distributed. We assume that the data is a random sample or random samples form a
normal population. The normal distribution has a probability density function which
is as below:
( ) ∞<<∞−
−
−= x
x
xf ,
2
1
exp.
1
.
2
1
2
σ
µ
σπ
.
Therefore , if we assume that a data set is normally distributed we imply that it has
this density function. Tests based on the assumption that the population has a certain
functional form are called parametric tests. Examples are t-tests and z-tests.
It is still possible to test hypotheses without assuming that the data has a certain
functional form. Tests conducted without the assumption of a certain functional form
are called nonparametric tests. The use of these is rather pragmatic because the
assumption that data has a certain distribution may not be valid sometimes. Besides
this, the sample size may not be big enought to justify the use the normal
approximation( Central Limit Theorem). Some nonparametric tests have higher
power to detect alternatives under some circumstances.
RANKING DATA
The table below (Table 1.1) shows measurements of blood pressure taken from 10
subjects. 4 received treatment and 6 were controls. Suppose we pool and arrange the
data in ascending order. Furthermore, if we assign numbers to these ranked values
and if the numbers assigned denote some ordering, we say that we are ranking the
data. The numbers which serve as the indices of location of the data points are called
ranks. Table 1.1 shows the data in the first two columns. Column 4 contains the
group labels and column 5 contains the ranks.
Table 1.1 : Measurements of Blood Pressure
Pressure Ranked Values of
Group (mmHg) Pressure Group Rank
1 94 78 2 1
1 108 80 2 2
1 110 85 2 3
1 90 88 2 4
2 80 90 1 5
2 85 94 1 6
2 94 94 2 7
2 78 105 2 8
2 105 108 1 9
2 88 110 1 10
3. iii
TIES IN A DATA SET
Suppose that any two values in a data set are equal. We call this a tie. When ties
occur, some remedial measures have to be taken. When several numbers are equal,
their rank is the average of their ranks. Consider the following example.
Measurements on some attributes were taken and are shown below:
94,108,110,90,80,94,85,90,90,90,108,94,78,105,88
Several values occur more than once. For instance 90, 94 and 108. The rank of each
90 is (5 + 6 + 7 + 8 )/4 = 6.5. The table below (Table 1.3) shows the ranked data.
Table 1.2: Ranked Data
Value 78 80 85 88 90 90 90 90 94 94 94 105 108 108 110
Rank 1 2 3 4 6.5 6.5 6.5 6.5 10 10 10 12 13.5 13.5 15
When ties occur in a data set, the test statistics calculated from that data have to be
adjusted to take care of this. This adjustment is called correction for ties. This is
sometimes hard to do manually. However, some statistical packages have in-built
mechanisms of doing this.
Some nonparametric tests are based on ranks. Statistical tests based on ranks are
called rank tests. Some parametric tests have nonparametric rank test analogues.
Examples are in the following table (Table 1.4).
Table 1.3 Nonparametric Analogues of Some Parametric Methods
Parametric Method Nonparametric Analogue
T-Test For Paired Samples Wilcoxon Signed Rank test, Sign Test
One Way Anova Kruskal Wallis Test
Two Way Anova Friedman Test
T-test for Independent Samples Rank Sum Test(Wilcoxon Mann-Whitney U
Test)
T-test for One Sample Sign Test
1. ONE SAMPLE TEST: THE SIGN TEST
These are tests based on data collected from one population. The parametric test one
can use is the t-test for small samples. If the sample size is large, one can use the z-
test. The sign test is based on the binomial distribution. Suppose we have a random
sample of observations from a certain population. Suppose also that our null
hypothesis is that the mean is equal to some constant, say µ0. Using statistical
notation, this can be written as Ho: µ = µ0. A strong assumption we will have to make
is that the distribution of the source population is symmetric so that the population
mean is equal to the median. Then, without having to assume that the data is
normally distributed, we can stage the same null hypothesis using the median as
follows :
4. iv
H0: M = M0, where M0 is the hypothesized median.
Let Di = Xi – M0, where Xi is observation i. Then each difference will be either
positive or negative. Let Yi be defined as follows:
Yi = 1, if Di > 0
= 0, if Di < 0
Let S be the sum of the Yi's. The probability of having a positive difference is equal
to the probability of having a negative difference. This is 0.5. If you take success as
a positive difference then S has a binomial distribution with parameters n and π=0.5.
Suppose that there are n0 ties. Then, there will be n0 difference equal to zero. Then S
has a binomial distribution with n-n0 and π = 0.5 as parameters. S is in essence, the
number of positive differences.
Example:
The following are measurements of weight, in kilograms, for 6 individuals:
62, 63, 64.5, 65, 72 and 60. We strongly believe that the mean is greater than 64 kg.
Find the deviations ( Di's ) and S, the sum of the Yi's.
Solution:
Table 1.4: Deviations From The Mean
Obs No(i) 1 2 3 4 5 6
Weight(kg) 62 63 64.5 65 72 60
Difference -2 -1 0.5 1 8 -4
Yi 0 0 1 1 1 1
Sign - - + + + +
Thus, S= 1+1+1+1 = 4. There are 4 positive differences.
S has a Binomial distribution with n = 6 and π = 0.5 as parameters. This is so because
the number of positive differences here is 6. π = 0.5 because the probability of having
a positive difference is just 0.5. We can thus use the binomial distribution to test the
following hypotheses:
1. H0 : M = M0 vs Ha : M ≠≠≠≠ M0
2. H0: M = M0 vs Ha : M > M0
3. H0: M = M0 vs Ha: M < M0
Implication of The Null Hypothesis
Under the null, the proportion of positive differences (π) is equal to the proportion of
negative differences. We thus assume that the data has symmetric distribution (not
necessarily bell-shaped). If a distribution is symmetric, µ =M.
Note that the null hypothesis has an equivalent expression and this is H0: π = 0.5.
Thus the above hypotheses can be rewritten as follows:
5. v
1. H0 : π = π0 vs Ha : π≠≠≠≠ π0
2. H0: π = π0 vs Ha : π> π0
3. H0: π = π0 vs Ha: π < π0
USE OF THE BINOMIAL DISTRIBUTION
To test the above hypotheses, we will use a cumulative binomial distribution. Recall
that the pmf of a binomial distribution is ( ) ( ) nr
r
n
rR
rnr
,...,1,0,1Pr =−
==
−
ππ .
The cumulative binomial distribution function is thus
( ) ( ) nr
r
n
rR
rnr
r
r
,,1,0,1Pr
0
0
0 Κ=−
=≤
−
=
∑ ππ
Suppose that we have a one-tail test (upper tail) and that our level of significance is
0.05. Then we can find which value of R gives us a cumulative probability of 0.95
approximately. The table below (Table 1.6) shows the cumulative probabilities of a
Bin(n = 6,π = 0.5) distribution.
Table 1.5:Bin(10, 0.5) Distribution
n R=r P Pr(R≤≤≤≤ r)
10 0 0.5 0.0010
10 1 0.5 0.0107
10 2 0.5 0.0547
10 3 0.5 0.1719
10 4 0.5 0.3770
10 5 0.5 0.6230
10 6 0.5 0.8281
10 7 0.5 0.9453
10 8 0.5 0.9893
10 9 0.5 0.9990
10 10 0.5 1.0000
From Table 1.6 above, R = 7 gives a probability of 0.95 approximately. This is our
cut-off (critical) point. We will therefore reject H0 if an observed value of R is greater
than 7. The critical region for our test is composed of 8,9 and 10.
Let us go back to our previous example about some measurements on weight in Table
1.5. There we have n = 6 and π = 0.5 (as usual). Our null hypothesis is H0 : π = 0.5
and our alternative is Ha: π > 0.5.
6. vi
The table below (Table 1.7) shows a Bin(6, 0.5) distribution.
Table 1.6: A Bin(6,0.5) Distribution
n R=r p Pr(R≤≤≤≤ r)
6 0 0.5 0.0156
6 1 0.5 0.1094
6 2 0.5 0.3438
6 3 0.5 0.6563
6 4 0.5 0.8906
6 5 0.5 0.9844
6 6 0.5 1.0000
We will test at the 2.5% level of significance. Our critical (cut-off) point is the value
of R which gives us a cumulative probability of, approximately, 0.975. R = 5 will do
because it gives us a cumulative probability of 0.9844 which is very close to 0.975.
We will therefore, reject H0 if R > 5 ( i.e. if R = 6). Here R = S, the sum of the Yi's.
From Table 1.5 above, S = 4 which implies that R = 4 for this data set. We accept H0
and conclude that the mean is equal to 64 kg. There are special tables for the
Cumulative Binomial Distribution. You can also use a computer to get the same
results. Some of the statistical packages which can be used are StatXact and SSPS.
Below is output from both StatXact and SPSS.
StatXact Output Panel
From the panel above, we see that the one-sided p-value is 0.0000. We therefore
reject Ho. The mean weight for this group is greater than 64 kg.
ESTIMATION OF BINOMIAL PARAMETER (PI)
Number of Trials =6
Number of Successes =4
Point Estimation of PI = 0.6667
97.50% Confidence Interval for PI = ( 0.1839 , 0.9699)
Exact P-values for testing PI = 0.0250
One-sided : Pr { T .GE. 4 } = 0.0000
Two-sided : 2 * One-sided = 0.0000
7. vii
SPSS Output Panel
Unfortunately, the SPSS panel is of no use because the p-value supplied is two-sided!
- - - - - Binomial Test
D
Cases
Test Prop. = .5000
4 LE .00 Obs. Prop. = .6667
2 GT .00
- Exact Binomial
6 Total 2-Tailed P = .6875
8. viii
2. TWO SAMPLE TESTS
2.1 INDEPENDENT SAMPLES TESTS
2.1.1 WILCOXON RANK SUM TEST/WILCOXON-MANN-WHITNEY
TEST
Suppose we are 'given' a group of N subjects to be used in a study. Let us take an
example of a clinical trial (an experimental study). If we want to compare a new drug
to a placebo1
, we will divide the subjects into two groups. These groups are called
arms of the trial. Arms of a trial may have different number of subjects. Let n denote
the number of subjects in the treatment group and m, the number of subjects in the
control group. Apparently, n + m = N.
The subjects are randomly allocated to the treatment and control groups. In some
trials, patients are not allowed to know which drug they are receiving. Only clinicians
in the study know. This administration of a drug without telling the patient which
drug he is getting is called blinding. The patient is 'blinded'. Since the clinician
knows the drug and the patient does not know what he is getting. Such blinding is
called single blinding. Single blinding eliminates the psychological effects that might
result if a patient knows which drug he is receiving. For example, a patient can fake
improvement! Sometimes both the clinician and the patient do not know which drug
is being administered. This is called double-blinding. This eliminates bias since the
clinician might curry favour if his relative or friend is in the study by giving him/her
the new drug being tested. Blood is thicker than water! Both the patient and the
clinician are blinded.
In this example, chance enters the study by the choice of subjects into the arms. We
can not claim that the group of N subjects is a sample from some population. This
scenario is very common to physicians. Here is a hospital ward. There are only 6
patients. Four are on Treatment A and two are on Treatment B. Measurements of
their blood pressure are taken. The main objective of the trial is to see if the two
treatments have the same effect on the blood pressure of patients. Here, the sample
size is very small. We can not even speak of asymptotics (Normal Approximation)!
The data may be skewed which may make it implausible to use the Student's t-test. A
very practical approach to this problem is by using exact nonparametric methods.
Since the selection of subjects into the two groups is done using random sampling, the
two samples can be said to be independent. There is a rank test which can be used to
compared the two groups. One such method is called the Wilcoxon Rank Sum Test.
Suppose that the responses for the two groups are pooled and ranked. Let Si denote
the rank of the ith
subject from the treatment group and Rj denote the rank of the jth
subject from the control group. Obviously, i = 1,…,n and j = 1,…,m. Our statistic is
the sum of ranks from the treatment groups. This is usually denoted by either Ws or
Tx. Similarly, the sum of the ranks for the control group is Wr. To carry out this test,
one must be very careful not to loose track of the group labels for the observations.
1
A placebo is a harmless pill containing no active ingredients
9. ix
The test statistic, Ws has a distribution and hypotheses can be tested using this
distribution. There are tables for this. This test is available in many statistical
packages e.g. StatXact.
DETERMINING THE ALTERNATIVE HYPOTHESIS.
Suppose that we are comparing two groups basing on measurements of weight in
kilograms. Group I was on Diet A and Group II was on Diet B. We want to know if
Diet A is more effective in increasing body weight than Diet B. We will therefore be
testing the following hypotheses:
H0: µ1 = µ2 vs Ha: µ1 > µ2
Higher values of weight in the first group will be in favour of Ha. Since we rank the
pooled weights, if Group I has higher weights on average than Group II then the rank
sum statistic, Ws will be very big. Diet A will be said to be more effective than Diet
B in increasing body weight if Ws ≥ c. The value of c is chosen so that Pr(Ws ≥ c) =
α.
Alternatively, one can test the same hypotheses using the significance probability.
Using some level of significance, α say, our significance probability is equal to Pr(Ws
≥ ws), where ws is the value of Ws calculated from the sample data. If we are testing
at α = 0.05, say, we can determine from special tables2
the value of Ws which gives us
this significance probability. This p-value is then compared to the level of
significance. Another name for 'significance probability' is p-value. The rest is
straightforward.
Example 1:
A mental hospital wishes to test the effectiveness of a new drug that is believed to
have a beneficial effect on some mental or emotional disorder. There are 5 patients in
the hospital suffering from this disorder. Three are selected at random to receive the
new drug and the other 2 serve as controls. The ranks of the treatment subjects are
2,3 and 5 and those of the control subjects are 1 and 4. Does the drug have a
beneficial effect? Test at α =0.05.
Solution:
i 1 2 3
Treatment ranks, Si 2 3 5
j 1 2
Control ranks, Rj 1 4
To carry out the hypothesis testing, the exact distribution of Ws must be known. This
can be worked out when N is small. For N = 5, the distribution of Ws is as below:
2
See the appendix
10. x
Table 2.1: The Distribution of Ws Using Permutation
ws P(Ws = ws) Pr(Ws ≤≤≤≤ ws)
6 0.100 0.100
7 0.100 0.200
8 0.200 0.400
9 0.200 0.600
10 0.200 0.800
11 0.100 0.900
12 0.100 1.000
The possible treatment ranks are permuted and the above distribution is arrived at.
Our observed value of Ws is ws = 2 + 3 + 5 = 10. From the above table (Table 2.1),
Pr(Ws ≥ 10) = 1 – Pr(Ws <10) = 1 - Pr(Ws ≤≤≤≤ 9) = 1 – 0.6 = 0.4. The p-value is 0.4. At
the 5% level of significance, we accept H0. The drug has no effect.
Most statistical packages have an equivalent form of the above test called the
Wilcoxon-Mann-Whitney Test or just Mann-Whitney Test.
USING THE NORMAL APPROXIMATION
When both m and n are less than 10, critical values and significance probabilities of
the Wilcoxon Rank Sum statistic can be obtained from the table in the appendix3
.
However, when n is greater than 10, we can use the normal approximation. To do
this, we need the mean and variance of Ws.
MEAN AND VARIANCE OF Ws
The mean of Ws, E(Ws) = )1(
2
1
+Nn ,
where N is the total number of subjects in the study and n is the number of subjects in
the treatment group.
The variance of Ws, Var(Ws) = ( )1
12
1
+Nmn ,
where m , n, and N are as before.
Similarly, E(Wr) = )1(
2
1
+Nm ,
where N is the total number of subjects in the study and m is the number of subjects in
the treatment group.
and
3
See the Appendix
11. xi
Var(Wr) = ( )1
12
1
+Nmn , where m , n, and N are as before.
Thus using the normal approximation, Z =
+
+−
)1(
12
1
)1(
2
1
Nmn
NnWs
has an approximate
standard normal distribution by the same Central Limit Theorem..
Similarly, Z =
+
+−
)1(
12
1
)1(
2
1
Nmn
NmWr
has an approximate standard normal distribution
by the same theorem. Therefore the p-value for a one-sided test,( Pr(Ws ≤≤≤≤ ws) or
Pr(Ws ≥ ws) ) can be calculated using this approximation.
Example2:
Suppose that m=10, n = 10 and Ws = 79. Calculate Pr(Ws ≤≤≤≤ 79) using the normal
approximation.
Solution:
E(Ws)= 105 and Var(Ws)=175
Thus Pr(Ws ≤≤≤≤ 79) =
( )
−
≤=
−
≤
23.13
26
Pr
175
10579
Pr ZZ
= ( ) .025.0965.1Pr =−≤Z
CORRECTION FOR TIES
Suppose that there are ties in a data set such that value number 1, in ascending order,
is occurring with frequency f1, value number two with frequency f2 and so on. Then
the mean of Ws* , E(Ws*) does not change. In other words, E(Ws) = E(Ws*).
However, the variance changes and it is equal to :
Var(Ws)* =
( )
( )
)1(1212
1 1
3
−
−
−
+
∑=
NN
ffmn
Nmn
e
i
ii
, where e is the frequency of the largest
observation.
12. xii
Example 3: Psychological counselling
In a test of the effect of psychological counselling, 80 boys are divided at random into
a control group of 40 to whom only the normal counselling facilities are available,
and a treatment group of 40 who receive special counselling. At the end of the study,
a careful assessment is made of each boy who is then classified as having made a
good, fairly good, fairly poor, or poor adjustment, with the following results.
Table 2.2 : Psychological Counselling Data
Poor Fairly Poor Fairly Good Good Total
Treatment 5 7 16 12 40
Control 7 9 15 9 40
Does counselling have a positive effect on adjustment? Test at α = 0.05.
Solution:
We will assign ranks to the five different categories of adjustment thus:
Poor = 1, Fairly Poor = 2, Fairly Good = 3, Good = 4. We can see from the above
table that we have ties in each category of adjustment. We have 5+7 = 12
observations tied at the rank 1, 16 at the rank 2, 31 at the rank 3, 21 and at the rank 4.
Below is a table of the ranks and the number of observations tied at each rank.
Table 2.3 Ranks And Number Of Observations Tied At Rank of Magnitude I
Rank 1 2 3 4
Number of observations
tied at rank i, fi
12 16 31 21
Average rank, Si* 6.5 20.5 44 70
These ranks are obtained as follows:
At the first position we have 12 observations tied. The mean of these ranks is
(1+ 2+…+12)/12 = 6.5. At the second position we have 16 numbers tied. The mean
rank is therefore (13+14+15+…+28)/16 = 20.5 and so on
Since our Wilcoxon Rank Sum statistic considers the treatment rank, Ws* = (6.5x12).
+(20.5 x 16) + (44 x 12) + (70 x 12) = 1,720. Since N = 80, m = n = 40 and f1 = 12,
f2 = 16, f3 = 31, f4 = 21 (which is also e=21), E(Ws* = 1,620) and Var(Ws*) = 99.27.
Our hypotheses are H0: Counselling has no positive effect
Ha: Counselling has a positive effect
Thus we will reject H0 if Pr(Ws* ≥ 1,720) < α = 0.05..
Now Pr(Ws* ≥ 1,720) = 1-Pr(Z ≤ 1.01)
= 0.16.
We accept H0 and conclude that counselling has no positive effect on adjustment.
13. xiii
2.12 MANN- WHITNEY TEST
Recall that Ws is the sum of treatment ranks. Let WXY = ( )1
2
1
+− nWs , where n is
defined as before. This statistic is the Mann-Whitney Statistic.
MEAN AND VARIANCE OF THE MANN WHITNEY STATISTIC
The mean of WXY, E(WXY) = mn
2
1
and its variance, Var (WXY) = ( )1
12
1
+Nmn .
CORRECTION FOR TIES
In the presence of ties, we use the same correction used for Ws*
. In such cases, the
mean of WXY
*
is equal to the mean of WXY when there are no ties. However, the
variance of WXY
*
is equal to the variance of Ws
*
.
USING THE NORMAL APPROXIMATION
Using the mean and variance of WXY
*
, p-value computations can be made using the
normal approximation.
Example 4:
1. Diastolic pressure was measured on 3 subjects in a treatment group and 4 subjects
in a control group. The data is displayed below in Table 1.1 below
Table 1.1: Blood Pressure of Subjects In A Clinical Trial
Treated
Group
Control Group
95 80
101 78
110 85
90
Does the drug have on effect on blood pressure? Test at α=0.05.
Solution:
(i) H0: The drug has no effect on blood pressure
H1: The drug has an effect on blood pressure
Note:
Since we do not say explicitly what effect the drug might have on pressure, this is a
two-tailed test.
14. xiv
(ii) Test statistic, Wilcoxon Rank Sum Test Statistic, Ws = ∑=
n
i
iS
1
where Si is the rank of subject I in the treatment group.
(iii) Pool the data and rank it. You can use Microsoft ExcelTM
to do this. When that
is done you get the results as presented in the following table, Table 2.4..
Table 2.4 Data On Blood Pressure Pooled And Ranked
Ranks for
Blood Pressure Group Rank Treatment Group
78 C 1
80 C 2
85 C 3
90 C 4
95 T 5 5
101 T 6 6
110 T 7 7
Sum 18
(iv) From the above table, the observed value of Ws is 18. (Ws = 18).
The Wilcoxon –Mann Whitney Statistic, WXY = ( )1
2
1
+− nWs
= 16)13(
2
1
18 =+−
(v) The significance probability (p-value ) for this is
Pr(|WXY| ≥ 16) = 2x Pr(WXY ≥ 16)
= 2x(1- Pr(WXY ≤ 16))
= 2x(1 – 1)
= 0.
Hence we reject Ho (since p-value < 0.05). The drug has an effect on the blood
pressure of subjects.
Using StatXact , one gets the following output,
15. xv
StatXact Output Panel
Explanation of The Above Output Panel
The mean is 12 which agrees with theory since
E(Ws) = ( ) ( ) 12173
2
1
1
2
1
=+=+ xxNn
The standard deviation is 2.828. We know that Variance of Ws,
Var(Ws) = ( ) 8834
12
1
1
12
1
==+ xxxNmn
Hence, the standard deviation of Ws, is = 828.28 =
The standardized value is 2.121. We know that using the normal approximation,
( )
( )
+
−
=
1
12
1
)(
Nmn
WsEWs
Z has an approximate normal distribution with parameters µ =0
and σ²= 1. ( i.e. Z ~ appr N(0, 1)). Hence, Z =
( ) 122.2
828.2
6
828.2
1218
==
−
StatXact gives approximate as well as exact p-values. When the sample size is large,
it is advisable to use the approximate p-value because then the normal approximation
holds ( Recall the Central Limit Theorem). In such cases of a very big sample size,
StatXact can not really use the exact method of computation to find the p-value. It is
not just feasible! Since our test is two-tailed, we will use the two-sided p-values.
From the approximate p-values in the above output panel, we reject Ho. This agrees
with what we found manually using Statistical Tables. It is appropriate in this case of
a small sample size to use exact methods.
WILCOXON-MANN-WHITNEY TEST
[ Sum of scores from population < 1 > ]
Summary of Exact distribution of WILCOXON-MANN-WHITNEY statistic:
Min Max Mean Std-dev Observed Standardized
6.000 18.00 12.00 2.828 18.00 2.121
Mann-Whitney Statistic = 12.00
Asymptotic Inference:
One-sided p-value: Pr { Test Statistic .GE. Observed } = 0.0169
Two-sided p-value: 2 * One-sided = 0.0339
Exact Inference:
One-sided p-value: Pr { Test Statistic .GE. Observed } = 0.0286
Pr { Test Statistic .EQ. Observed } = 0.0286
Two-sided p-value: Pr { | Test Statistic - Mean |
.GE. | Observed - Mean | = 0.0571
Two-sided p-value: 2*One-Sided = 0.0571
16. xvi
Using exact p-values from the same output panel, we accept Ho; the treatment has no
effect.
Example 4: (Example 2 continued)
We found the mean and standard deviation of WXY
*
to be 1,620 and 99.29
respectively. The mean and standard deviation of WXY* are 180 and 99.29
respectively. Hence, Pr(WXY
*
1720) = Pr(Z = 1.007) = 0.16 as before. The
conclusion does not change. We still accept H0.
PAIRED SAMPLES T TEST.
1. THE SIGN TEST
In section, we encountered the sign test for the first time as a nonparametric test of
location for one sample. Suppose that we have two paired samples. Let Xi denote the
response (outcome) to some treatment A for subject i in the first group and Yi the
response of subject i in the second group to treatment B.
Using the paired samples t-test, we can test whether treatment A is better than
treatment B. We use the paired differences between the paired measurements Xi and
Yi. Let Di = Yi-Xi. Regardless of the size of the difference, the differences will be
either positive or negative. Our interest is in the positive differences. If Treatment A
is indeed better than treatment B, then, on average, we expect Di to be positive. As
before, define Yi as below:
Yi = 1, if Di > 0
= 0, if Di < 0
The sum of the Yi's, S say, will be the total number of positive differences which is
our statistic. S has a binomial distribution with the sample size and p = 0.5 as
parameters. When there are ties, the difference, Di will be equal to zero. One simple
solution to this is to reduce n by the number of zeros. For instance, if there are 6 zeror
and the sample size is 20, the value of n to use in testing will be 14.
DISTRIBUTION OF S
Since S has a binomial distribution with N and π = 0.5,
The mean of S, E(S) =
2
n
n =π and its variance , Var(S) =
44
1
2
1
.)1(
n
xxnn ==−ππ .
If n is large, one can use the normal approximation.
Example:
The following data from Makutch and Parks (1988) document the response of
serum antigen level to AZT in 20 AIDS patients. Two sets of antigen levels are
provided for each patient; pre-treatment and post-treatment. The differences are also
displayed, along with two sets of signed ranks.
17. xvii
Table:2.5 AZT and Serum Antigen Trial
Sign
Patient Serum Antigen Level(pg/ml) Test
Id Pre-AZT Post-AZT Difference Scores
1 149 0 -149 0
2 0 51 51 1
3 0 0 0 -
4 259 385 126 1
5 106 0 -106 0
6 255 235 -20 0
7 0 0 0 -
8 52 0 -52 0
9 340 48 -392 0
10 65 65 0 -
11 180 77 -103 0
12 0 0 0 -
13 84 0 -84 0
14 89 0 -89 0
15 212 53 -159 0
16 554 150 -404 0
17 500 0 -500 0
18 424 165 -259 0
19 112 98 -14 0
20 2600 0 -2600 0
Sum 2
Does AZT have an effect on serum antigen levels? Test at α = 0.05.
Solution:
There are 20 differences, 4 of which are zeros. Therefore, for our sign test n = 16.
Two differences are positive. Therefore S=2. Calculation of a p-value will depend on
our alternative hypothesis. If our alternative is that AZT increases serum antigen
level then our p-value will be calculated as follows:
∑∑ =
=
≤−===≥=
0
1
0
16
2
)Pr(1)Pr()2Pr(
s
s
s
sSsSSp
= ( ) ( ( ) ( ) ( ) ( ) )151160
5.05.0
1
16
5.5.0
0
16
1)1Pr()0Pr(1
+
−==+=− SS
= 1-(0.00001526 + 0.0002441)
= 0.000259
Hence we reject H0. AZT increases serum antigen levels in AIDS patients.
Using StatXact, we get the following output in the panel below.
18. xviii
Explanation Of Output:
Manually, we found S equal to 2. From the above output,under the column 'Observed'
we have 2 which is the observed value of S. Our alternative is non-directional since
we do not say what effect AZT has. Therefore, we will use the exact two-sided p-
value which is equal to 0.0042. We reject H0. AZT has an effect on Serum Antigen
Levels in AIDS patients.
The Sign Test for paired samples is good and simple to use. However, there is loss of
information since the actual magnitude of the differences is not taken into
consideration. As such, it as has a power which is less than that of a test which takes
into consideration the actual size of the differences. On such a method is the
Wilcoxon Signed Rank Test.
2. THE WILCOXON SIGNED RANK TEST
This method is for paired samples and is similar to the Sign Test for paired
differences. The only minor difference is that the magnitude of the differences are
considered in testing. The method is as follows:
1. Find the difference, Di between paired values.
1. Ignoring the sign of the difference ( i.e. whether positive or negative) arrange the
differences in ascending order.
2. For each difference,attach the sign of the difference to the rank of that difference.
3. Add all ranks with a positive sign. Let the sum of these positive ranks be TWS.
This is your statistic.
SIGN TEST
Summary of Exact distribution of SIGN statistic:
Min Max Mean Std-dev Observed Standardized
0.0000 16.00 8.000 2.000 2.000 -3.000
Asymptotic Inference:
One-sided p-value: Pr { Test Statistic .LE. Observed } = 0.0013
Two-sided p-value: 2 * One-sided = 0.0027
Exact Inference:
One-sided p-value: Pr { Test Statistic .LE. Observed } = 0.0021
Pr { Test Statistic .EQ. Observed } = 0.0018
Two-sided p-value: 2*One-Sided = 0.0042
19. xix
THE DISTRIBUTION OF TWS.
The mean of Tws, E(Tws) =
( )
4
1+NN
and its variance , Var(Tws) =
4
N
. One can use
these two for the normal approximation.
Example:
We will revisit the AZT example. The data is reproduced below with another column
added.
Patient Signed
Id Pre-AZT Post AZT Difference Ranks
1 149 0 -149 -10
2 0 51 51 3
3 0 0 0 .
4 259 385 126 9
5 106 0 -106 -8
6 255 235 -20 -2
7 0 0 0 .
8 52 0 -52 -4
9 340 48 -392 -13
10 65 65 0 .
11 180 77 -103 -7
12 0 0 0 .
13 84 0 -84 -5
14 89 0 -89 -6
15 212 53 -159 -11
16 554 150 -404 -14
17 500 0 -500 -15
18 424 165 -259 -12
19 112 98 -14 -1
20 2600 0 -2600 -16
Sum 12
Column 5 contains the ranks with signs attached to them and the total of positive
ranks is 9 + 3 =12.
Using StatXact:
On the next is an output panel from StatXact.
20. xx
Since our test is two-tailed we will use the two-sided p-value. This p-value is equal to
0.0021 which is less than 0.05, our level of significance. We still reject H0. AZT has
an effect on Serum Antigen levels in AIDS patients. Note that the observed value is
given in the output panel under the column observed as 12, just what we have
calculated at the bottom of column 5 of the table above.
GENERAL NOTE:
WHEN TO USE TWO SIDED ALTERNATIVES.
1. When trying to decide which of two treatments is better and both are new, there is
no reason to stage an alternative hypothesis which says that one is better than the
other.
2. If the problem is not deciding which of two treatments or procedures is better but
that one wants to know whether they differ at all then the alternative must be two-
sided. For example, in a study comparing a surgical approach to a medical
problem with a more conservative treatment, it may be desirable to examine first
whether it is necessary to distinguish between two surgical techniques used by
different surgeons.
WILCOXON SIGNED RANK TEST
Summary of Exact distribution of WILCOXON SIGNED RANK statistic:
Min Max Mean Std-dev Observed Standardized
0.0000 136.0 68.00 19.34 12.00 -2.896
Asymptotic Inference:
One-sided p-value: Pr { Test Statistic .LE. Observed } = 0.0019
Two-sided p-value: 2 * One-sided = 0.0038
Exact Inference:
One-sided p-value: Pr { Test Statistic .LE. Observed } = 0.0011
Pr { Test Statistic .EQ. Observed } = 0.0002
Two-sided p-value: Pr { | Test Statistic - Mean |
.GE. | Observed - Mean | = 0.0021
Two-sided p-value: 2*One-Sided = 0.0021
21. xxi
EXERCISES
1. From a group of nine rats available for a study of learning, five were selected at
random and were trained to imitate the leader rat in a maze. They were then
placed together with four untrained rats in a situation where imitation of the
leaders enabled them to avoid receiving an electric shock. The results (the
number of trials required to obtain ten correct responses in ten consecutive trials)
were as follows:
Trained rats 74 64 75 45 82
Control 110 70 53 51
Find the significance probability of these results when the Wilcoxon Rank Sum Test
is used. What do you conclude? (Test at α = 0.05).
2. The effectiveness of vitamin C in orange juice and in synthetic ascorbic acid was
compared in 20 guinea pigs (divided at random into two groups of 10) in terms of
the length of the odontoblasts after 6 weeks, with the following results:
Orange
juice
8.2 9.4 9.6 9.7 10.0 14.5 15.2 16.1 17.6 21.5
Ascorbic
acid
4.2 5.2 5.8 6.4 7.0 7.3 10.1 11.2 11.3 11.5
Test the hypothesis of no difference against the hypothesis that the orange juice tends
to give larger values. Use α = 0.05.
(a) Use the Rank sum test
(b) Sign Test
(c) Wilcoxocon Signed Rank Test.
3. Suppose that a new postsurgical treatment is being compared with a standard
treatment by observing the recovry times of 9 treatment subjects and 9 controls.
The data is in the table below:
Standard
Treatment
20 21 24 30 32 36 40 48 54
New
Treatment
19 22 25 26 28 29 34 37 38
(a) Find the p-value if the Wilcoxon Rank Sum Test is used.
(b) Using the Wilcoxon Signed Rank Test to test the null hypothesis that the new
treatment is no better than the standard one. Use α = 0.05.
4. Suppose that 20 treatment patients are being compared with 20 controls, and that
the progress of each patient is classified as very poor, poor, indifferent, good, or
very good. The data are given in the following table:
22. xxii
Very Poor Poor Indifferent Good Very Good
Control 2 2 11 4 1
Treatment 0 1 9 7 3
Is the treatment effective? Test at α = 0.05.
5. In an investigation of a new drug for postoperative pain relief, it was desired to
determine (among other things) whether the relief brought by 3 mg of the drug is
significntly higher than that resulting from a dose of 1 mg. In one phase of the
study, 15 freshly operated-upon patients were assigned at random, 7 to the lower
dose (T1) and the remaining 8 to the higher dose (T3). The responses (number of
hours of pain relief) were T1: 2,0,3,3,0,0,8 and T3: 6,4,4,0,1,8,2,8. How
significant are these results? Use any appropriate nonparametric test. (α = 0.05 ).
23. xxiii
REFERENCES
Lehmann, E.L (1975). Nonparametrics: Statistical Methods Based on Ranks.San
Fransisco:Holden-Day Inc.
Gajjar,Y.,Mehta, C., Patel, N., Senchaudhuri, P., StatXact: User's Manual.
24. xxiv
PREFACE
It is not very easy to produce a text which is strictly non-technical. Of course, when
designing course material the audience has to be at heart. How user-friendly a text is
is very subjective and it depends on ones background. I have tried to make the stuff
attractive, user-friendly and straight-forward without losing too much rigour. I have
also tried my best to make it simple. I could not have made it any simpler. After all,
Albert Enstein said that we can make things simple but not simpler!
Nonparametric statistics is seen as difficult to some because they are not exposed to it
at an early stage. However, it is very useful in biomedical research and is like any
other statistical method. Its importance in inference can not be over-emphasized.
The methods explained in this brochure can be tried out using some statistical
packages. One such a statistical package is StatXact. This is chiefly for
nonparametric statistics. Other statistical packages also have routines for carrying
out non parametric inferential procedures. An example in mind is SPSS.
A reader who is looking for more rigorous stuff should read the books mentioned in
the bilbiography.
Any comments, corrections and necessary additions should be addressed to the
author. Constructive criticisms are welcome.
I hope you will enjoy this brochure.
Blantyre,
14 December, 2000
H. E. Misiri
HEMisiri@yahoo.co.uk
25. xxv
TABLE OF CONTENTS
Page
Preface ii
Introduction 1
Ranking Data 1
Ties In A Data Set 2
Nonparametric Analogues of Some Parametric Methods 2
Sign Test for One Sample 2
Using The Binomial Distribution 4
Two sample Tests 7
Wilcoxon Rank Sum test 7
Determining The Alternative 8
Using The Normal Approximation to The Distribution of Ws 9
Correcting for Ties When using Ws 10
Wilcoxon –Mann-Whitney Test 12
Nonparametric Rank - Based Paired Samples Tests 15
The Wilcoxon Signed Rank Test 17
The Distribution of The Wilcoxon Signed Rank Test 18
When To use Two-Sided Alternatives 19
Exercises 20
Bilbiography 22
Appendix 23