This document discusses correlation analysis, including definitions, types of correlation (positive, negative, linear, nonlinear), and methods of studying correlation (scatter diagrams, correlation graphs, coefficient of correlation). Correlation refers to the relationship between two or more variables, where they either move in the same direction (positive correlation) or opposite directions (negative correlation). Correlation does not necessarily imply causation. Methods of measuring correlation include scatter diagrams, correlation graphs, and Karl Pearson's coefficient of correlation which provides a numerical measurement of the correlation between -1 and 1.
This document discusses correlation and provides examples of its applications. It begins with an introduction that defines correlation as measuring the linear relationship between two variables. It then provides definitions of positive and negative correlation. The next sections discuss types of correlation based on degree, number of variables, and linearity. Correlation coefficient is introduced as a measure of the strength of the linear relationship between -1 and 1. Examples of its applications include the relationships between tree cutting and erosion, study time and test scores, clothing size and growth. Limitations of only considering linear relationships are also covered. Real-life examples of positive, negative, and no correlation between variables like temperature and sales, exercise and body fat, and weather and sales are presented.
This document discusses correlation and different aspects of studying correlation. It defines correlation as the association or relationship between two variables that do not cause each other. It describes different types of correlation including positive, negative, linear, non-linear, simple, multiple and partial correlation. It also discusses various methods of studying correlation including graphic methods like scattered diagrams and correlation graphs, and algebraic methods like Karl Pearson's correlation coefficient and Spearman's rank correlation coefficient. The document explains concepts like coefficient of determination and hypothesis testing in correlation. It emphasizes that correlation indicates association but does not necessarily imply causation between variables.
The document discusses the concept of correlation, specifically linear correlation. It provides definitions of correlation from various sources and explains that correlation refers to the relationship between two or more variables. The degree of this relationship is measured by the correlation coefficient. Common types of correlation are discussed such as positive and negative correlation. Methods for studying correlation are also outlined, including scatter diagrams and Karl Pearson's coefficient of correlation.
This document defines correlation and discusses different types of correlation. It states that correlation refers to the relationship between two variables, where their values change together. There can be positive correlation, where variables change in the same direction, or negative correlation, where they change in opposite directions. Correlation can also be linear, nonlinear, simple, multiple, or partial. The degree of correlation is measured by the coefficient of correlation, which ranges from -1 to 1. Graphic and algebraic methods like scatter diagrams and calculating the coefficient can be used to study correlation.
The document presents a presentation on coefficient correlation by Irshad Narejo. It defines correlation as a technique used to measure the relationship between two or more variables. A correlation coefficient measures the degree to which changes in one variable can predict changes in another, though correlation does not imply causation. Correlation coefficient formulas return a value between -1 and 1 to indicate the strength and direction of relationships between data. Positive correlation means high values in one variable are associated with high values in the other, while negative correlation means high values in one variable are associated with low values in the other. The document discusses Pearson's correlation coefficient formula and provides an example of calculating correlation by hand versus using SPSS.
This document discusses correlation analysis in agriculture. It begins by defining correlation as the relationship between two or more variables. Some key points:
- Correlation can be positive (variables move in the same direction), negative (variables move in opposite directions), linear, nonlinear, simple, multiple, partial or total.
- Common types analyzed in agriculture include the relationship between yield and rainfall, price and supply, height and weight.
- Methods for measuring correlation are discussed, including Karl Pearson's coefficient of correlation (denoted by r), Spearman's rank correlation, and scatter diagrams.
- The value of r ranges from -1 to 1, with higher positive or negative values indicating a stronger linear relationship between variables
This document discusses correlation analysis, including definitions, types of correlation (positive, negative, linear, nonlinear), and methods of studying correlation (scatter diagrams, correlation graphs, coefficient of correlation). Correlation refers to the relationship between two or more variables, where they either move in the same direction (positive correlation) or opposite directions (negative correlation). Correlation does not necessarily imply causation. Methods of measuring correlation include scatter diagrams, correlation graphs, and Karl Pearson's coefficient of correlation which provides a numerical measurement of the correlation between -1 and 1.
This document discusses correlation and provides examples of its applications. It begins with an introduction that defines correlation as measuring the linear relationship between two variables. It then provides definitions of positive and negative correlation. The next sections discuss types of correlation based on degree, number of variables, and linearity. Correlation coefficient is introduced as a measure of the strength of the linear relationship between -1 and 1. Examples of its applications include the relationships between tree cutting and erosion, study time and test scores, clothing size and growth. Limitations of only considering linear relationships are also covered. Real-life examples of positive, negative, and no correlation between variables like temperature and sales, exercise and body fat, and weather and sales are presented.
This document discusses correlation and different aspects of studying correlation. It defines correlation as the association or relationship between two variables that do not cause each other. It describes different types of correlation including positive, negative, linear, non-linear, simple, multiple and partial correlation. It also discusses various methods of studying correlation including graphic methods like scattered diagrams and correlation graphs, and algebraic methods like Karl Pearson's correlation coefficient and Spearman's rank correlation coefficient. The document explains concepts like coefficient of determination and hypothesis testing in correlation. It emphasizes that correlation indicates association but does not necessarily imply causation between variables.
The document discusses the concept of correlation, specifically linear correlation. It provides definitions of correlation from various sources and explains that correlation refers to the relationship between two or more variables. The degree of this relationship is measured by the correlation coefficient. Common types of correlation are discussed such as positive and negative correlation. Methods for studying correlation are also outlined, including scatter diagrams and Karl Pearson's coefficient of correlation.
This document defines correlation and discusses different types of correlation. It states that correlation refers to the relationship between two variables, where their values change together. There can be positive correlation, where variables change in the same direction, or negative correlation, where they change in opposite directions. Correlation can also be linear, nonlinear, simple, multiple, or partial. The degree of correlation is measured by the coefficient of correlation, which ranges from -1 to 1. Graphic and algebraic methods like scatter diagrams and calculating the coefficient can be used to study correlation.
The document presents a presentation on coefficient correlation by Irshad Narejo. It defines correlation as a technique used to measure the relationship between two or more variables. A correlation coefficient measures the degree to which changes in one variable can predict changes in another, though correlation does not imply causation. Correlation coefficient formulas return a value between -1 and 1 to indicate the strength and direction of relationships between data. Positive correlation means high values in one variable are associated with high values in the other, while negative correlation means high values in one variable are associated with low values in the other. The document discusses Pearson's correlation coefficient formula and provides an example of calculating correlation by hand versus using SPSS.
This document discusses correlation analysis in agriculture. It begins by defining correlation as the relationship between two or more variables. Some key points:
- Correlation can be positive (variables move in the same direction), negative (variables move in opposite directions), linear, nonlinear, simple, multiple, partial or total.
- Common types analyzed in agriculture include the relationship between yield and rainfall, price and supply, height and weight.
- Methods for measuring correlation are discussed, including Karl Pearson's coefficient of correlation (denoted by r), Spearman's rank correlation, and scatter diagrams.
- The value of r ranges from -1 to 1, with higher positive or negative values indicating a stronger linear relationship between variables
This document discusses correlation analysis and its various types. Correlation is the degree of relationship between two or more variables. There are three stages to solve correlation problems: determining the relationship, measuring significance, and establishing causation. Correlation can be positive, negative, simple, partial, or multiple depending on the direction and number of variables. It is used to understand relationships, reduce uncertainty in predictions, and present average relationships. Conditions like probable error and coefficient of determination help interpret correlation values.
Correlation analysis is used to find the degree of relationship between two or more variables by applying statistical tools. It produces a correlation coefficient that describes the strength and direction of the relationship. There are different types of correlation, including positive correlation where variables move in the same direction, negative correlation where they move in opposite directions, simple correlation between two variables, partial correlation controlling for other variables, and multiple correlation between three or more variables. Correlation analysis is important for measuring the degree of relationship between variables, estimating their values, and understanding economic behavior.
This is about the correlation analysis in statistics. It covers types, importance,Scatter diagram method
Karl pearson correlation coefficient
Spearman rank correlation coefficient
This document discusses regression, comparing it to causation and correlation. Regression analysis estimates and predicts the average value of one variable based on the values of other variables. For example, predicting a son's average height from his father's height. Causation indicates a relationship where changing one variable affects another. Correlation measures the association between variables, while regression numerically relates an independent variable to a dependent variable to estimate or predict values.
Correlation analysis is used to determine the relationship between two or more variables. It can analyze the degree, direction, and type of relationship. The key types of correlation are positive (variables increase together), negative (variables change inversely), simple (two variables), partial (three+ variables with some held constant), and multiple (three+ variables together). Correlation can also be linear (constant ratio of changes) or non-linear (varying ratio of changes). It is useful for understanding variable behavior, estimating values, and interpreting results with measures like the correlation coefficient and coefficient of determination.
The document discusses different types of correlation and methods for studying correlation. It describes Karl Pearson's coefficient of correlation, which measures the strength and direction of a linear relationship between two variables. The coefficient ranges from -1 to 1, where -1 is a perfect negative correlation, 0 is no correlation, and 1 is a perfect positive correlation. The document also discusses other types of correlation coefficients like Spearman's rank correlation coefficient and methods for analyzing correlation like scatter plots.
- Francis Galton carried out early work on correlation and his colleagues Karl Pearson developed methods to calculate correlation coefficients for parametric data.
- Correlation describes the relationship between two variables on a scale from -1 to 1, where -1 is a perfect negative correlation, 0 is no correlation, and 1 is a perfect positive correlation.
- Pearson's correlation coefficient r is commonly used to determine the strength and direction of correlation between two variables, with values between -0.6 and -1 or 0.6 and 1 indicating strong correlation.
- Correlation is widely used in research in physical education to analyze relationships between factors like physical performance and strength.
Correlation and regression analysis are statistical tools used to analyze relationships between variables. Correlation measures the strength and direction of association between two variables on a scale from -1 to 1. Regression analysis uses one variable to predict the value of another variable and draws a best-fit line to represent their relationship. There are always two lines of regression - one showing the regression of x on y and the other showing the regression of y on x. Regression coefficients from these lines indicate the slope and intercept of the lines and can help estimate unknown variable values based on known values.
1. The document discusses correlation and correlation coefficients, which measure the strength and direction of association between two variables.
2. A correlation coefficient ranges from 0, indicating no correlation, to 1 or -1, indicating perfect positive or negative correlation. Coefficients above 0.5 generally indicate a strong linear relationship.
3. The Pearson correlation coefficient (r) specifically measures the linear correlation between two normally distributed variables, while the Spearman correlation (rs) is nonparametric and assesses correlation between ordinal or non-normally distributed variables.
4. Correlation only indicates association, not causation. Significant correlation is also not necessarily clinically meaningful. Correlation coefficients and their statistical significance must be interpreted carefully.
The document discusses correlation and regression analysis. It defines positive and negative correlation, as well as linear and non-linear correlation. It provides examples of variables that are positively and negatively correlated. It also discusses how correlation coefficients measure the strength of the relationship between two variables from -1 to 1. Regression analysis uses regression equations to predict unknown variable values from known variable values.
This document discusses correlation analysis and different types of correlation. It defines correlation as the degree of inter-relatedness between two or more variables. Correlation can be positive, negative, simple, partial or multiple depending on the direction and number of variables analyzed. Linear and non-linear correlation are determined based on whether the relationship between variables is constant or not. The document provides examples of calculating correlation coefficients in Excel and SPSS and computing partial correlation coefficients controlling for other variables.
Fundamental of Statistics and Types of CorrelationsRajesh Verma
This document provides an overview of key concepts in statistics including parametric vs non-parametric statistics, descriptive vs inferential statistics, types of errors, significance levels, correlation, and different correlation coefficients. Parametric statistics rely on assumptions of the normal distribution while non-parametric do not. Descriptive statistics describe data and inferential statistics draw conclusions. Type I and II errors occur when the null hypothesis is incorrectly rejected or not rejected. Significance levels like 0.05 are used to determine statistical significance. Correlation measures the relationship between variables from -1 to 1. Different coefficients like Pearson, Spearman, and Kendall's Tau are used depending on the scale of measurement and data distribution.
Covariance is a measure of how two random variables change together, taking any value from -∞ to +∞. Covariance can be affected by changing the units of the variables. Correlation is a scaled version of covariance that indicates the strength of the relationship between two variables on a scale of -1 to 1. Unlike covariance, correlation is not affected by changes in the location or scale of the variables and provides a standardized measure of their relationship. Correlation is therefore preferred over covariance as a measure of the relationship between two variables.
The document discusses covariance and correlation, which describe the relationship between two variables. Covariance indicates whether variables are positively or inversely related, while correlation also measures the degree of their relationship. A positive covariance/correlation means variables move in the same direction, while a negative covariance/correlation means they move in opposite directions. Correlation coefficients range from 1 to -1, with 1 indicating a perfect positive correlation and -1 a perfect inverse correlation. The document provides formulas for calculating covariance and correlation and examples to demonstrate their use.
Statistical techniques for measuring the closeness of the relationship between variables.It measures the degree to which changes in one variable are associated with changes in another.It can only indicate the degree of association or covariance between variables. Covariance is a measure of the extent to which two variables are related.
Correlation- an introduction and application of spearman rank correlation by...Gunjan Verma
this presentation contains the types of correlation, uses, limitations, introduction to spearman rank correlation, and its application. a numerical is also given in the presentation
This document is a presentation by Dwaiti Roy on partial correlation. It begins with an acknowledgement section thanking various professors and resources that helped in preparing the presentation. It then provides definitions and explanations of key concepts related to partial correlation such as correlation, assumptions of correlation, coefficient of correlation, coefficient of determination, variates, partial correlation, assumptions and hypothesis of partial correlation, order and formula of partial correlation. Examples are provided to illustrate partial correlation. The document concludes with references and suggestions for further reading.
Correlation analysis examines the relationship between two or more variables. Positive correlation means the variables increase together, while negative correlation means they change in opposite directions. The Pearson correlation coefficient, r, quantifies the strength of linear correlation between -1 and 1. Multiple correlation analysis extends this to measure the correlation between one dependent variable and multiple independent variables. It is useful but assumes linear relationships and can be complex to calculate.
The document discusses correlation and the conditions that must be checked before calculating correlation between two quantitative variables. It explains that correlation measures the strength of the linear association between variables and can only be applied to quantitative, not categorical, variables. Additionally, the relationship must be sufficiently linear for correlation to be valid and outliers can distort results. The document also defines positive and negative correlation and different types of correlation based on the number of variables.
The document discusses correlation analysis and different types of correlation. It defines correlation as the linear association between two random variables. There are three main types of correlation:
1) Positive vs negative vs no correlation based on the relationship between two variables as one increases or decreases.
2) Linear vs non-linear correlation based on the shape of the relationship when plotted on a graph.
3) Simple vs multiple vs partial correlation based on the number of variables.
The document also discusses methods for studying correlation including scatter plots, Karl Pearson's coefficient of correlation r, and Spearman's rank correlation coefficient. It provides interpretations of the correlation coefficient r and coefficient of determination r2.
The document discusses how the output of processes with multiple factors often follows a normal distribution. It provides examples using dice to illustrate how the distribution patterns change as more dice (factors) are added, moving from a uniform to triangular to bell-shaped normal distribution. The key points covered are:
- Processes with multiple random factors that act simultaneously will often result in a normal distribution for the total output or effect.
- The normal distribution is symmetrical and bell-shaped. Nearly all (99.73%) of the population lies within 3 standard deviations of the mean.
- Real-life processes have many random factors that combine to produce outputs that tend to be normally distributed.
This document discusses correlation analysis and its various types. Correlation is the degree of relationship between two or more variables. There are three stages to solve correlation problems: determining the relationship, measuring significance, and establishing causation. Correlation can be positive, negative, simple, partial, or multiple depending on the direction and number of variables. It is used to understand relationships, reduce uncertainty in predictions, and present average relationships. Conditions like probable error and coefficient of determination help interpret correlation values.
Correlation analysis is used to find the degree of relationship between two or more variables by applying statistical tools. It produces a correlation coefficient that describes the strength and direction of the relationship. There are different types of correlation, including positive correlation where variables move in the same direction, negative correlation where they move in opposite directions, simple correlation between two variables, partial correlation controlling for other variables, and multiple correlation between three or more variables. Correlation analysis is important for measuring the degree of relationship between variables, estimating their values, and understanding economic behavior.
This is about the correlation analysis in statistics. It covers types, importance,Scatter diagram method
Karl pearson correlation coefficient
Spearman rank correlation coefficient
This document discusses regression, comparing it to causation and correlation. Regression analysis estimates and predicts the average value of one variable based on the values of other variables. For example, predicting a son's average height from his father's height. Causation indicates a relationship where changing one variable affects another. Correlation measures the association between variables, while regression numerically relates an independent variable to a dependent variable to estimate or predict values.
Correlation analysis is used to determine the relationship between two or more variables. It can analyze the degree, direction, and type of relationship. The key types of correlation are positive (variables increase together), negative (variables change inversely), simple (two variables), partial (three+ variables with some held constant), and multiple (three+ variables together). Correlation can also be linear (constant ratio of changes) or non-linear (varying ratio of changes). It is useful for understanding variable behavior, estimating values, and interpreting results with measures like the correlation coefficient and coefficient of determination.
The document discusses different types of correlation and methods for studying correlation. It describes Karl Pearson's coefficient of correlation, which measures the strength and direction of a linear relationship between two variables. The coefficient ranges from -1 to 1, where -1 is a perfect negative correlation, 0 is no correlation, and 1 is a perfect positive correlation. The document also discusses other types of correlation coefficients like Spearman's rank correlation coefficient and methods for analyzing correlation like scatter plots.
- Francis Galton carried out early work on correlation and his colleagues Karl Pearson developed methods to calculate correlation coefficients for parametric data.
- Correlation describes the relationship between two variables on a scale from -1 to 1, where -1 is a perfect negative correlation, 0 is no correlation, and 1 is a perfect positive correlation.
- Pearson's correlation coefficient r is commonly used to determine the strength and direction of correlation between two variables, with values between -0.6 and -1 or 0.6 and 1 indicating strong correlation.
- Correlation is widely used in research in physical education to analyze relationships between factors like physical performance and strength.
Correlation and regression analysis are statistical tools used to analyze relationships between variables. Correlation measures the strength and direction of association between two variables on a scale from -1 to 1. Regression analysis uses one variable to predict the value of another variable and draws a best-fit line to represent their relationship. There are always two lines of regression - one showing the regression of x on y and the other showing the regression of y on x. Regression coefficients from these lines indicate the slope and intercept of the lines and can help estimate unknown variable values based on known values.
1. The document discusses correlation and correlation coefficients, which measure the strength and direction of association between two variables.
2. A correlation coefficient ranges from 0, indicating no correlation, to 1 or -1, indicating perfect positive or negative correlation. Coefficients above 0.5 generally indicate a strong linear relationship.
3. The Pearson correlation coefficient (r) specifically measures the linear correlation between two normally distributed variables, while the Spearman correlation (rs) is nonparametric and assesses correlation between ordinal or non-normally distributed variables.
4. Correlation only indicates association, not causation. Significant correlation is also not necessarily clinically meaningful. Correlation coefficients and their statistical significance must be interpreted carefully.
The document discusses correlation and regression analysis. It defines positive and negative correlation, as well as linear and non-linear correlation. It provides examples of variables that are positively and negatively correlated. It also discusses how correlation coefficients measure the strength of the relationship between two variables from -1 to 1. Regression analysis uses regression equations to predict unknown variable values from known variable values.
This document discusses correlation analysis and different types of correlation. It defines correlation as the degree of inter-relatedness between two or more variables. Correlation can be positive, negative, simple, partial or multiple depending on the direction and number of variables analyzed. Linear and non-linear correlation are determined based on whether the relationship between variables is constant or not. The document provides examples of calculating correlation coefficients in Excel and SPSS and computing partial correlation coefficients controlling for other variables.
Fundamental of Statistics and Types of CorrelationsRajesh Verma
This document provides an overview of key concepts in statistics including parametric vs non-parametric statistics, descriptive vs inferential statistics, types of errors, significance levels, correlation, and different correlation coefficients. Parametric statistics rely on assumptions of the normal distribution while non-parametric do not. Descriptive statistics describe data and inferential statistics draw conclusions. Type I and II errors occur when the null hypothesis is incorrectly rejected or not rejected. Significance levels like 0.05 are used to determine statistical significance. Correlation measures the relationship between variables from -1 to 1. Different coefficients like Pearson, Spearman, and Kendall's Tau are used depending on the scale of measurement and data distribution.
Covariance is a measure of how two random variables change together, taking any value from -∞ to +∞. Covariance can be affected by changing the units of the variables. Correlation is a scaled version of covariance that indicates the strength of the relationship between two variables on a scale of -1 to 1. Unlike covariance, correlation is not affected by changes in the location or scale of the variables and provides a standardized measure of their relationship. Correlation is therefore preferred over covariance as a measure of the relationship between two variables.
The document discusses covariance and correlation, which describe the relationship between two variables. Covariance indicates whether variables are positively or inversely related, while correlation also measures the degree of their relationship. A positive covariance/correlation means variables move in the same direction, while a negative covariance/correlation means they move in opposite directions. Correlation coefficients range from 1 to -1, with 1 indicating a perfect positive correlation and -1 a perfect inverse correlation. The document provides formulas for calculating covariance and correlation and examples to demonstrate their use.
Statistical techniques for measuring the closeness of the relationship between variables.It measures the degree to which changes in one variable are associated with changes in another.It can only indicate the degree of association or covariance between variables. Covariance is a measure of the extent to which two variables are related.
Correlation- an introduction and application of spearman rank correlation by...Gunjan Verma
this presentation contains the types of correlation, uses, limitations, introduction to spearman rank correlation, and its application. a numerical is also given in the presentation
This document is a presentation by Dwaiti Roy on partial correlation. It begins with an acknowledgement section thanking various professors and resources that helped in preparing the presentation. It then provides definitions and explanations of key concepts related to partial correlation such as correlation, assumptions of correlation, coefficient of correlation, coefficient of determination, variates, partial correlation, assumptions and hypothesis of partial correlation, order and formula of partial correlation. Examples are provided to illustrate partial correlation. The document concludes with references and suggestions for further reading.
Correlation analysis examines the relationship between two or more variables. Positive correlation means the variables increase together, while negative correlation means they change in opposite directions. The Pearson correlation coefficient, r, quantifies the strength of linear correlation between -1 and 1. Multiple correlation analysis extends this to measure the correlation between one dependent variable and multiple independent variables. It is useful but assumes linear relationships and can be complex to calculate.
The document discusses correlation and the conditions that must be checked before calculating correlation between two quantitative variables. It explains that correlation measures the strength of the linear association between variables and can only be applied to quantitative, not categorical, variables. Additionally, the relationship must be sufficiently linear for correlation to be valid and outliers can distort results. The document also defines positive and negative correlation and different types of correlation based on the number of variables.
The document discusses correlation analysis and different types of correlation. It defines correlation as the linear association between two random variables. There are three main types of correlation:
1) Positive vs negative vs no correlation based on the relationship between two variables as one increases or decreases.
2) Linear vs non-linear correlation based on the shape of the relationship when plotted on a graph.
3) Simple vs multiple vs partial correlation based on the number of variables.
The document also discusses methods for studying correlation including scatter plots, Karl Pearson's coefficient of correlation r, and Spearman's rank correlation coefficient. It provides interpretations of the correlation coefficient r and coefficient of determination r2.
The document discusses how the output of processes with multiple factors often follows a normal distribution. It provides examples using dice to illustrate how the distribution patterns change as more dice (factors) are added, moving from a uniform to triangular to bell-shaped normal distribution. The key points covered are:
- Processes with multiple random factors that act simultaneously will often result in a normal distribution for the total output or effect.
- The normal distribution is symmetrical and bell-shaped. Nearly all (99.73%) of the population lies within 3 standard deviations of the mean.
- Real-life processes have many random factors that combine to produce outputs that tend to be normally distributed.
The document discusses the parties, settings, purposes, and reference sources for psychological testing and assessment. It identifies the main parties as test developers/publishers, test users, test takers, and society. Assessments are commonly conducted in educational, geriatric, counseling, clinical, business, military, and other settings to evaluate individuals' abilities, diagnose issues, inform treatment, and make organizational decisions. Authoritative information on specific tests can be found in test manuals, reference books, journal articles, online databases, and test publishers' catalogs.
The document discusses psychological research methods. It begins by defining research and its goals, which include describing behavior, establishing relationships between causes and effects, and developing theories about human behavior. It then describes the empirical research cycle and different research methods, both primary like experiments and secondary like meta-analyses. It discusses variables, research designs, qualitative and quantitative data collection and analysis, and drawing conclusions. Finally, it covers ethical issues in research and challenges in determining causality.
This document discusses methods for analyzing relationships between categorical variables using two-way tables. It explains how to calculate marginal distributions by finding row totals, column totals, and grand totals from two-way tables. Conditional distributions are also discussed, which show the proportion of each category within each row or column total. An example two-way table is provided analyzing the relationship between age group and gender using percentages to describe the marginal and conditional distributions.
This document provides an introduction to psychological testing and assessment. It defines key terms like tests, testing, and assessment. It describes the different types of constructs that can be measured by psychological tests, like traits, conditions, intelligence, and attitudes. It also discusses the differences between testing and assessment and contrasts the two processes. The document outlines the various tools used in psychological assessment, including tests, interviews, observations, and more. It discusses important topics like test administration and interpretation, the parties involved in testing, and the various settings where assessment is conducted. Finally, it covers issues like culture, ethics, and laws related to psychological assessment.
Psychological testing is a field characterized by the use of samples of performance in order to assess psychological construct, such as cognitive and emotional implementation, about a given individual.
Psychological tests are formal tools used to measure mental functioning and behaviors. They can be administered in various settings like schools, hospitals, and workplaces to assess abilities, personality, and neurological status. Common uses of tests include education placement, career counseling, diagnosing disorders, and selecting job applicants. Tests vary in their administration method, targeted behaviors, and purpose between ability, personality, and clinical domains. Proper tests are standardized, objective, use norms, and are reliable and valid measures of their intended construct.
Correlation is a statistical technique used to determine the degree of relationship between two variables. Correlational research aims to identify and describe relationships but does not imply causation. Positive correlation indicates high scores on one variable are associated with high scores on the other, while negative correlation means high scores on one variable are associated with low scores on the other. Correlational research can be used for explanatory or predictive purposes. More complex techniques like multiple regression allow prediction using combinations of variables. Threats to internal validity like subject characteristics must be controlled.
This document discusses psychological assessment and tests. It describes the development and types of psychological tests, including intelligence tests like the Stanford-Binet and Wechsler scales, achievement tests, aptitude tests, personality tests like the MMPI, and projective tests like the Rorschach inkblot test. It also outlines the nurse's role in psychological assessment, which includes educating patients, observing behaviors, and documenting changes.
Psychological tests were developed to assist in understanding human behavior and making important decisions in an objective manner. Tests provide standardized samples of behavior that can be used to infer underlying traits and make comparisons to norms. This allows for decisions to be made with less bias than relying solely on subjective human judgment. Tests quantify results to precisely describe behaviors and allow for clearer communication than qualitative descriptions alone.
This document discusses rank correlation and Spearman's rank correlation coefficient. It defines correlation as a relationship between two variables where a change in one variable corresponds to a change in the other. Rank correlation involves ranking observations from highest to lowest rather than using the original values, which avoids assumptions about the population distribution. Spearman's rank correlation coefficient measures the correspondence between two rankings and is calculated based on the differences between ranks of paired items. It provides a distribution-free measure of correlation.
Correlation of subjects in school (b.ed notes)Namrata Saxena
This document discusses the concept of correlation in education. It defines correlation as the mutual relationship between different subjects or variables in a curriculum. The document outlines the importance of correlation, including that it helps students perceive knowledge as a whole, strengthens retention of knowledge, and promotes well-rounded development. It discusses different types of correlation, including vertical/internal correlation between topics within a subject and horizontal/external correlation between different subjects. Examples are provided of how mathematics can be correlated with other subjects like science, geography, and economics.
Psychological test meaning, concept, need & importancejd singh
This document discusses psychological testing. It defines psychological testing as a standardized measure of a person's behavior that is used to observe differences among individuals. It notes that tests measure constructs like abilities, functioning, and personality. The document outlines the objectives, need, importance and types of psychological tests. It describes the major characteristics of tests including standardization, norms, reliability and validity. Finally, it provides examples of commonly used psychological tests.
Data organization and presentation (statistics for research)Harve Abella
The document discusses various methods of presenting data, including textual, tabular, and graphical displays. It provides examples and definitions of key terms used in data presentation, such as frequency distribution tables, class intervals, class boundaries, class marks, and cumulative frequencies. The document also outlines steps for constructing a frequency distribution table, including determining the number of classes, range, class size, and class limits.
The document discusses the normal distribution and its key properties: bell-shaped and symmetrical around the mean, extending from negative to positive infinity with an area under the curve of 1. Approximately 95% and 99.9% of the distribution lies within 2 and 3 standard deviations of the mean, respectively. It also discusses how to calculate probabilities using the standard normal distribution where the mean is 0 and standard deviation is 1, and how to standardize other normal distributions.
Correlation describes the relationship between two or more variables. A positive correlation means that as one variable increases, the other also increases, while a negative correlation means that as one variable increases, the other decreases. Correlation is measured numerically using coefficients like the Pearson correlation coefficient r, which ranges from -1 to 1, with values farther from 0 indicating stronger linear relationships and the direction indicating positive or negative correlation. Correlation is used in business and economics to study relationships between variables like price and demand.
Point biserial correlation measures the relationship between one dichotomous variable (with two possible values) and one continuous variable. It ranges from -1 to +1. A positive correlation indicates that as the dichotomous variable increases in value, so does the continuous variable, and vice versa. An example is measuring the correlation between depression status (depressed or not depressed) and shame scores (continuous values from 1-10). The direction of the correlation depends on how the variables are coded.
This document provides an introduction to correlation and regression analysis. It defines correlation as a measure of the association between two variables and regression as using one variable to predict another. The key aspects covered are:
- Calculating correlation using Pearson's correlation coefficient r to measure the strength and direction of association between variables.
- Performing simple linear regression to find the "line of best fit" to predict a dependent variable from an independent variable.
- Using a TI-83 calculator to graphically display scatter plots of data and calculate the regression equation and correlation coefficient.
A Guide to SlideShare Analytics - Excerpts from Hubspot's Step by Step Guide ...SlideShare
This document provides a summary of the analytics available through SlideShare for monitoring the performance of presentations. It outlines the key metrics that can be viewed such as total views, actions, and traffic sources over different time periods. The analytics help users identify topics and presentation styles that resonate best with audiences based on view and engagement numbers. They also allow users to calculate important metrics like view-to-contact conversion rates. Regular review of the analytics insights helps users improve future presentations and marketing strategies.
Correlation and Regression analysis is one of the important concepts of statistics which could be used to understand the relationship between the variables.
Module - 2 correlation and regression.pptxjayvee73
1. Correlation analyzes the relationship between two variables, determining the degree to which they change together. Regression finds the relationship between a dependent and independent variable to estimate or predict future values.
2. Simple linear regression analyzes the relationship between one dependent and one independent variable. Multiple linear regression generalizes this to multiple independent variables.
3. Correlation is measured using methods like scatter diagrams, Pearson's correlation coefficient, and Spearman's rank correlation coefficient. The degree of correlation ranges from perfect to zero.
Module - 2 correlation and regression.pptxjayvee73
1. Correlation analyzes the relationship between two variables, determining the degree to which they change together. Regression finds the relationship between a dependent and independent variable to estimate or predict future values.
2. Simple linear regression analyzes the relationship between one dependent and one independent variable. Multiple linear regression generalizes this to multiple independent variables.
3. Correlation is measured using methods like scatter diagrams, Pearson's correlation coefficient, and Spearman's rank correlation coefficient. The degree of correlation ranges from perfect to zero.
This document defines correlation and discusses different types of correlation. It states that correlation refers to the relationship between two variables, where their values change together. There can be positive correlation, where variables change in the same direction, or negative correlation, where they change in opposite directions. Correlation can also be linear, nonlinear, simple, multiple, or partial. The degree of correlation is measured using a coefficient of correlation between -1 and 1, indicating no, perfect, or limited correlation. Correlation is studied using scatter diagrams or graphs and calculated using formulas to find the coefficient.
correlation or dependence is any statistical relationship, whether causal or not, between two random variables or bivariate data. Although in the broadest sense, "correlation" may indicate any type of association, in statistics it normally refers to the degree to which a pair of variables are linearly related. Familiar examples of dependent phenomena include the correlation between the height of parents and their offspring, and the correlation between the price of a good and the quantity the consumers are willing to purchase, as it is depicted in the so-called demand curve.
Correlations are useful because they can indicate a predictive relationship that can be exploited in practice. For example, an electrical utility may produce less power on a mild day based on the correlation between electricity demand and weather. In this example, there is a causal relationship, because extreme weather causes people to use more electricity for heating or cooling. However, in general, the presence of a correlation is not sufficient to infer the presence of a causal relationship (i.e., correlation does not imply causation).
Formally, random variables are dependent if they do not satisfy a mathematical property of probabilistic independence. In informal parlance, correlation is synonymous with dependence. However, when used in a technical sense, correlation refers to any of several specific types of mathematical operations between the tested variables and their respective expected values. Essentially, correlation is the measure of how two or more variables are related to one another. There are several correlation coefficients, often denoted
ρ
\rho or
r
r, measuring the degree of correlation. The most common of these is the Pearson correlation coefficient, which is sensitive only to a linear relationship between two variables (which may be present even when one variable is a nonlinear function of the other). Other correlation coefficients – such as Spearman's rank correlation – have been developed to be more robust than Pearson's, that is, more sensitive to nonlinear relationships.[1][2][3] Mutual information can also be applied to measure dependence between two variables.
This document discusses correlation analysis and its various types. Correlation is a measure of the relationship between two or more variables. There are three main types of correlation based on the degree, number of variables, and linearity. Correlation can be positive, negative, simple, partial, multiple, linear, or non-linear. Correlation is important for understanding relationships between variables, making predictions, and interpreting data. However, correlation does not necessarily imply causation.
This document discusses correlation analysis and different types of correlation. It defines correlation as a statistical analysis of the relationship between two or more variables. There are three main types of correlation discussed:
1. Positive correlation means that as one variable increases, the other also tends to increase. Negative correlation means that as one variable increases, the other tends to decrease.
2. Simple correlation analyzes the relationship between two variables, while multiple correlation analyzes three or more variables simultaneously. Partial correlation holds the effect of other variables constant.
3. Methods for measuring correlation include scatter diagrams, which graphically show the relationship, and algebraic formulas that calculate a correlation coefficient to quantify the strength and direction of the relationship.
The document discusses correlation and regression analysis. It defines correlation as the statistical relationship between two variables, where a change in one variable corresponds to a change in the other. The key types of correlation are positive, negative, simple, partial and multiple, and linear and non-linear. Regression analysis establishes the average relationship between an independent and dependent variable in order to predict or estimate values of the dependent variable based on the independent variable. Methods for studying correlation include scatter diagrams and Karl Pearson's coefficient of correlation, while regression analysis uses equations to model the linear relationship between variables.
Correlation analysis measures the strength and direction of association between two or more variables. It is represented by the coefficient of correlation (r), which ranges from -1 to 1. A value of 0 indicates no association, 1 indicates perfect positive association, and -1 indicates perfect negative association. The scatter diagram is a graphical method to visualize the association between variables by plotting their values. Karl Pearson's coefficient is a commonly used algebraic method to calculate the coefficient of correlation from sample data.
Multivariate Analysis Degree of association between two variable- Test of Ho...NiezelPertimos
The document discusses multivariate analysis and correlation. It defines correlation as a measure of the degree of association between two variables. A correlation coefficient between -1 and 1 indicates the strength and direction of the linear relationship, with values closer to 1 or -1 being stronger. Positive correlation means the variables move in the same direction, while negative correlation means they move in opposite directions. The document provides examples and methods for calculating and interpreting correlation coefficients, including using scatter plots and the Pearson product-moment formula. Excel functions for finding correlation across multiple data sets are also described.
Correlation analysis is a statistical method used to measure the strength of the linear relationship between two variables. A high correlation indicates a strong relationship, while a low correlation means the variables are weakly related. Researchers use correlation analysis in market research to identify relationships, patterns, and trends between variables. There are three types of correlation - positive, negative, and no correlation. Methods for studying correlation include scatter diagrams and Karl Pearson's coefficient of correlation. Spearman's rank correlation coefficient is used when variables are qualitative rather than quantitative.
Correlation and regression are statistical techniques used to analyze relationships between variables. Correlation determines the strength and direction of a relationship, while regression describes the linear relationship to predict changes in one variable based on changes in another. There are different types of correlation including simple, multiple, and partial correlation. Regression analysis determines the regression line that best fits the data to estimate values of one variable based on the other. The correlation coefficient measures the strength of linear correlation from -1 to 1, while regression coefficients are used to predict changes in the variables.
Correlation refers to a statistical relationship between two variables where they change together, but does not imply that one causes the other. Causation means one event causes the other. It is difficult to establish causation compared to correlation. Correlation is measured using a correlation coefficient from +1 to -1 indicating the strength and direction of the relationship. While correlation coefficients can detect linear relationships between variables, they do not prove causation. Controlled studies that compare outcomes of groups receiving different treatments can help establish causality.
This document discusses correlation analysis and the different types of correlation. It begins by defining correlation as the degree of relationship between two or more variables. Correlation can be positive, negative, or zero. Positive correlation means that as one variable increases, the other also increases. Negative correlation means that as one variable increases, the other decreases. Zero correlation means there is no relationship between the variables. Correlation can also be linear or non-linear. Linear correlation follows a perfect straight line pattern, while non-linear correlation follows a curved pattern. The document also briefly discusses methods for studying correlation such as scatter diagrams and Pearson's correlation coefficient.
This document discusses correlation analysis. It defines correlation as the degree of relationship between two random variables. Correlation can be positive, negative, simple, partial, or multiple depending on the direction and number of variables. Linear correlation means changes in one variable are proportionally related to changes in the other. Non-linear correlation means changes are not proportional. Correlation is measured using the correlation coefficient r, which ranges from -1 to 1. A higher absolute r value means stronger correlation. Correlation only indicates relationship and not causation. The document also covers probable error and coefficient of determination in interpreting correlation results.
The management of a regional bus line thought the companys cost of .pdfnagaraj138348
The management of a regional bus line thought the company\'s cost of gas might be correlated
with its passenger/mile ratio. The data and a correlation matrix follow Comment.
Solution
In statistics, dependence refers to any statistical relationship between two random
variables or two sets of data. Correlation refers to any of a broad class of statistical relationships
involving dependence. Familiar examples of dependent phenomena include the correlation
between the physical statures of parents and their offspring, and the correlation between the
demand for a product and its price. Correlations are useful because they can indicate a predictive
relationship that can be exploited in practice. For example, an electrical utility may produce less
power on a mild day based on the correlation between electricity demand and weather. In this
example there is a causal relationship, because extreme weather causes people to use more
electricity for heating or cooling; however, statistical dependence is not sufficient to demonstrate
the presence of such a causal relationship (i.e., Correlation does not imply causation). Formally,
dependence refers to any situation in which random variables do not satisfy a mathematical
condition of probabilistic independence. In loose usage, correlation can refer to any departure of
two or more random variables from independence, but technically it refers to any of several more
specialized types of relationship between mean values. There are several correlation coefficients,
often denoted ? or r, measuring the degree of correlation. The most common of these is the
Pearson correlation coefficient, which is sensitive only to a linear relationship between two
variables (which may exist even if one is a nonlinear function of the other). Other correlation
coefficients have been developed to be more robust than the Pearson correlation – that is, more
sensitive to nonlinear relationships.[1][2][3] Several sets of (x, y) points, with the Pearson
correlation coefficient of x and y for each set. Note that the correlation reflects the noisiness and
direction of a linear relationship (top row), but not the slope of that relationship (middle), nor
many aspects of nonlinear relationships (bottom). N.B.: the figure in the center has a slope of 0
but in that case the correlation coefficient is undefined because the variance of Y is zero.
Contents [show] [edit]Pearson\'s product-moment coefficient Main article: Pearson product-
moment correlation coefficient The most familiar measure of dependence between two quantities
is the Pearson product-moment correlation coefficient, or \"Pearson\'s correlation.\" It is obtained
by dividing the covariance of the two variables by the product of their standard deviations. Karl
Pearson developed the coefficient from a similar but slightly different idea by Francis Galton.[4]
The population correlation coefficient ?X,Y between two random variables X and Y with
expected values µX and µY and standard devia.
This document discusses correlation and linear regression. It defines correlation as the statistical relationship between two variables, ranging from -1 to 1. Positive correlation means the variables increase or decrease together, while negative correlation means they deviate in opposite directions. Linear regression analyzes the linear relationship between a dependent and independent variable to predict future outcomes. It allows executives to forecast sales, understand how variables influence each other, and prepare budgets based on regression equations.
This document contains a final term project report submitted by three students to their professor. The report summarizes statistical techniques including correlation, regression, and measures of central tendency. For correlation, the document defines correlation, describes the correlation coefficient and different types of correlation. It also discusses the history and uses of correlation. For regression, it defines regression, describes the regression coefficient and line, and discusses the history and uses of regression. Finally, it defines different measures of central tendency including mean, median, mode, and discusses their advantages and disadvantages. The report is presented in a table of contents and contains examples, formulas and multiple choice questions.
The presentation covers topics like Investment and Speculation, Investment and Gambling, Investment Management Process, Types of Speculators, Technical Analysis and Fundamental Analysis, Concept of Risk and Return
THE INVESTMENT ENVIRONMENT - PART 1: Meaning of Investment/Types of Investments/Characteristics of Investment/Objectives of Investment/Types of Investors/Investment Management Process
Tabulation involves arranging classified data in a systematic table with rows and columns. An effective table has a number, title, captions, stubs, body, head/foot notes, and source. The table example shows student details from Irshadiya College classified by gender, family, and residence type. There are general rules for preparing clear and useful tables, such as using a rough sketch, ordering elements alphabetically or chronologically, maintaining uniformity, avoiding excessive information, and distinguishing important entries. Tabulation simplifies complex data, facilitates comparison, presents data in proper perspective, and reveals patterns.
Single entry system of accounting is on of the easiest methods of preparing financial statements. This presentation discuss the various aspects of Single Entry System of Accounting
it is the era of advertising. We are going through a number of advertisements in our day to day to day life. In the period of technological development, especially internet and other technologies, advertisements will also have to change a lot. This presentation will help you to understand the various aspects of Electronic Advertisements
Internet marketing is a form of e-marketing that uses internet technologies to achieve marketing objectives. It provides advantages like more freedom and information for customers, speedy customization, and lower costs. However, it also has disadvantages such as an inability to digitize some products, a lack of touch experience for customers, and potential misuse of customer information. Emerging trends in internet marketing include growth, expanding to B2B segments, advergaming, digital media, relationship marketing, and mobile and rich media access.
E-marketing refers to marketing goods and services through digital media and electronic devices. It involves building and maintaining customer relationships through electronic means like the internet to facilitate exchanges that help achieve marketing objectives. Some key benefits of e-marketing include lower transaction costs, an ability to immediately respond to customers, and effective customer targeting. However, e-marketing also faces challenges like potential for misleading information, high setup costs, and risks to customer privacy.
A review of the growth of the Israel Genealogy Research Association Database Collection for the last 12 months. Our collection is now passed the 3 million mark and still growing. See which archives have contributed the most. See the different types of records we have, and which years have had records added. You can also see what we have for the future.
हिंदी वर्णमाला पीपीटी, hindi alphabet PPT presentation, hindi varnamala PPT, Hindi Varnamala pdf, हिंदी स्वर, हिंदी व्यंजन, sikhiye hindi varnmala, dr. mulla adam ali, hindi language and literature, hindi alphabet with drawing, hindi alphabet pdf, hindi varnamala for childrens, hindi language, hindi varnamala practice for kids, https://www.drmullaadamali.com
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
This presentation was provided by Steph Pollock of The American Psychological Association’s Journals Program, and Damita Snow, of The American Society of Civil Engineers (ASCE), for the initial session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session One: 'Setting Expectations: a DEIA Primer,' was held June 6, 2024.
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
How to Make a Field Mandatory in Odoo 17Celine George
In Odoo, making a field required can be done through both Python code and XML views. When you set the required attribute to True in Python code, it makes the field required across all views where it's used. Conversely, when you set the required attribute in XML views, it makes the field required only in the context of that particular view.
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
1. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 1
Study Notes on
PSYCHOLOGICAL STATISTICS
SEMESTER-III
B.Sc. COUNSELING PSYCHOLOGY
Prepared by:
Noushad P.K
Lecturer in Commerce
Irshadiya College of Commerce and Social Sciences, Feroke
Department of Psychology,
Irshadiya College of Commerce and Social Sciences
Feroke, Calicut-6736 31
MODULE –I
CORRELEATION ANALYSIS
MEANING AND DEFINITION
It is a statistical tool used to describe the relationship or interdependence
between two or more variables. Two or more variables are said to be correlated, if
the change in one variable results in a corresponding change in the other variable.
That is, when two or more variables move together (in same direction or in
opposite direction), we say they are correlated. For instance, when the price of a
commodity increases, its supply goes up. On the basis of the theory of correlation,
one can study the comparative changes occurring in two related phenomena and
their cause-effect relationship.
A.M Tuttle defines “Correlation as an analysis of the association between
two or more variables”. In the words of Simpson and Kafka “Correlation analysis
deals with the association between two or more variables”.
SIGNIFICANCE OF CORRELATION ANALYSIS
1. It helps to find a single figure to measure the relationship between the
variables.
2. It helps to understand the economic behaviour.
3. It can be used as a basis for the study of regression
4. It helps to reduce the range of uncertainty associated with decision
making
5. It helps to know whether the correlation is significant or not. This is
possible by comparing the correlation co efficient with 6PE. If ‘r’ is more
than 6 PE, the correlation is significant.
CORRELATION AND CAUSATION
For correlation, it is usually implies that the two variables should have
cause-effect relationship. For example, the relationship between price and
demand, price and supply, rate of interest and savings, etc.
Correlation does not always imply cause-effect relationship. For example,
a higher degree of correlation between yield per acre of rice and tea may be due
to the fact that both are related to the amount of rainfall.
There may be a higher degree of correlation between the variables, but it
may be difficult to pinpoint as to which is the effect. For example, increase in price
leads to decrease in demand. Here change in price is the cause and change in
demand is the effect. But it is also possible that increased demand is due to other
reasons like growth of population.
2. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 2
Two series showing high degree of correlation may be purely from chance
also. For example, during the last decade there has been a significant increase in
the sale of newspaper and crime. We can establish correlation between these two
variables. But there exists no cause-effect relationship between these two factors.
Such illogical correlations are known as Non sensical Correlation/Spurious
Correlation
CLASSIFICATION OF CORRELATION
Positive and Negative Correlation
Correlation can be either positive or negative. Whether correlation is
positive or negative would depend upon the direction in which the variables are
moving.
When the value of two variables move in the same direction, correlation
is said to be positive. That is, an increase in the value of one variable results an
increase in the value other variable also, or, a decrease in the value of one
variable leads to a decrease in other variable also. Example, correlation between
price and supply
Price: 10 20 30 40 50
Supply: 80 100 150 170 200
When the value of two variables move in the opposite direction,
correlation is said to be negative. That is, an increase in the value of one variable
results a decrease in the value of other variable. Example, correlation between
price and demand
Price: 5 10 15 20 25
Demand: 16 10 8 6 2
Linear and Non-linear Correlation
Correlation may be linear or non-linear. Here the distinction is based upon
the constancy of the ratio of change between the variables.
When the amount of change in one variable leads to a constant ratio of
change in the other variable, correlation is said to be linear. In a correlation
analysis, if the ratio of change between the two sets of variable is same, then it is
called linear correlation. In such a case, if the values of variables are plotted on a
graph paper, then a straight line is obtained. Example, if price goes up by 10%, it
leads to a rise in supply by 15% each time
Price: 10 15 30 60
Supply: 50 75 150 300
When the amount of change in one variable does not bring the same ratio
of change in the other variable, that is the ratio happens to be variable instead of
constant, the correlation is said to be non-linear. In such a case, we shall obtain a
curve, if the values of variables are plotted on a graph paper. That is why it is
known as curvy-linear correlation. For example,
X: 2 4 6 10 15
Y: 8 10 18 22 26
Simple, Partial and Multiple Correlations
In a correlation analysis, if only two variables are studied (of which one is
independent and the other is dependent), the correlation is said to be simple. For
example, the correlation between price and demand.
Multiple correlation studies the relationship between a dependent
variable and two or more independent variables. For example, the correlation
between yield with both rainfall and temperature
In partial correlation, we measure the correlation between a dependent
variable and one particular independent variable assuming that all other
independent variables remain constant. For example, there are three variables-
yield, rainfall and temperature. And each is related with the other. Then, the
relationship between yield and rainfall (assuming the temperature is constant) is
the partial correlation
METHODS OF STUDYING CORRELATION
1. Graphic method
a) Scatter diagram
b) Correlation graph
2. Algebraic methods/Mathematical methods/statistical methods/Co-
efficient of correlation methods
a) Karl Pearson’s Co-efficient of correlation
b) Spearman’s Rank correlation method
c) Concurrent deviation method
SCATTER DIAGRAM
It is also known as dot chart. It is a graphical method of studying
correlation between two variables. It is a visual aid to show the presence or
absence of correlation between two variables.
In scatter diagram, one of the variables is shown on the X-axis and the
other on Y-axis. Each pair of values is plotted by means of a dot mark. If these dot
marks show some trends either upward or downward, the two variables are said
3. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 3
to be correlated. If the plotted dots do not show any trend, the two variables are
not correlated. The greater the scatter of the dots, the lower is the relationship
Merits of Scatter Diagram Method
1. It is a simple method.
2. It is a non-mathematical method.
3. It is very easy to understand.
4. It is not affected by the size of extreme values.
5. It is usually the first step in correlation analysis.
Demerits of Scatter Diagram Method
1. It gives only a rough idea about the correlation between variables.
2. Further algebraic treatment is not possible.
3. The exact degree of correlation between the variables cannot be easily
determined.
4. If the number of pairs of variables is either very big or very small, the
method is not easy.
CORRELATION GRAPH METHOD
Under correlation graph method the individual values of the two variables
are plotted on a graph paper. Then dots relating to these variables are joined
separately so as to get two curves. By examining the direction and closeness of the
two curves, we can infer whether the variables are related or not. If both the
curves are moving in the same direction (either upward or downward) correlation
is said to be positive. If the curves are moving in the opposite directions,
correlation is said to be negative.
Merits of Correlation Graph Method
1. It is a simple method.
2. It does not require mathematical calculations.
3. It is very easy to understand
Demerits of Correlation Graph Method:
1. A numerical value of correlation cannot be calculated.
2. It is only a pictorial presentation of the relationship between variables.
3. It is not possible to establish the exact degree of relationship between the
variables.
MATHEMATICAL/STATISTICAL CORRELATION/CO-EFFICIENT OF
CORRELATION
It is an algebraic method of measuring correlation. It measures the degree
or extent of correlation between two variables. It is symbolically denoted by r.
The value of correlation co-efficient can never be +1 or -1. That is, +1 and -1 are
the limits of this co-efficient. It covers:
1. Karl Pearson’s co-efficient of correlation
2. Spearman’s rank correlation
3. Concurrent deviation
Karl Pearson’s Co-Efficient of Correlation
It was developed by the reputed statistician and biologist Prof: Karl
Pearson. It is also known as product moment correlation co-efficient.
Assumptions
1. There is a linear relationship between variables.
2. The variables are affected by a large number of dependent causes so as to
form a normal distribution.
3. There is a cause-effect relationship between the variables.
Properties
1. It has a well-defined formula.
2. It is a pure number and is independent of the units of measurement.
3. It lies in between ±1
4. It is the geometric mean of the two regressions co-efficient.
5. It does not change with reference to change of origin or change of scales.
6. Co-efficient of correlation between x and y is same as that between y
and x
Methods
a) When deviations are taken from actual mean
b) When deviations are taken from assumed mean
When deviations are taken from actual mean
STEPS:
1. Take the deviations of x series from the mean of x which is denoted by x
or dx
2. Square these deviations and get total. That is, ∑ or ∑ .
4. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 4
3. Take the deviations of y series from the mean of y which is denoted by y
or dy
4. Square these deviations and get total. That is, ∑ or ∑ .
5. Multiply the deviations of x and y series, and get the total. That is
∑
6. Apply the formula and find correlation co-efficient.
∑
√∑ ∑
Where, x = ̅
Y= ̅
OR
∑
√∑ ∑
Where, dx = ̅
dy= ̅
Illustration-1:
Find Karl Pearson’s co-efficient of correlation from the following data.
X: 8 4 10 2 6
Y: 9 11 5 8 7
Solution:
X Y
x
=
( ̅)
Y
=
( ̅)
x2
y2
xy
8
4
10
2
6
9
11
5
8
7
2
-2
4
-4
0
1
3
-3
0
-1
4
4
16
16
0
1
9
9
0
1
2
-6
-12
0
0
∑ ∑ ∑ = 40 ∑ = 20 ∑
Arithmetic Mean of X, ̅ =
∑
= = 6
Arithmetic Mean of Y, ̅ =
∑
= = 8
Correlation,
∑
√∑ ∑
=
√
=
√
= = - 0.565
Illustration-2:
Calculate Karl Pearson’s co-efficient of correlation, from the following data?
X: 2 3 4 5 6 7 8
Y: 4 5 6 12 9 5 4
Solution:
X Y
dx =
( ̅)
dy =
( ̅)
dx
2
dy
2
dx.dy
2
3
4
5
6
7
8
4
5
6
12
9
7
6
-3
-2
-1
0
1
2
3
-3
-2
-1
5
2
0
-1
9
4
1
0
1
4
9
9
4
1
25
4
0
1
9
4
1
0
2
0
-3
∑ ∑ ∑ = 28 ∑ = 44 ∑
Arithmetic Mean of X, ̅ =
∑
= = 5
Arithmetic Mean of Y, ̅ =
∑
= = 7
Correlation,
∑
√∑ ∑
=
√
=
√
= = 0.37
When deviations are taken from assumed mean
STEPS:
1. Take the deviations of x series from the assumed mean of x which is
denoted by dx
2. Square these deviations and get total. That is, ∑ .
3. Take the deviations of y series from the assumed mean of y which is
denoted by dy
4. Square these deviations and get total. That is, ∑ .
5. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 5
5. Multiply the deviations of x and y series, and get the total. That is
∑
6. Apply the formula and find correlation co-efficient.
∑
(∑ )(∑ )
√∑
(∑ ) √∑
(∑ )
Where, dx = X- Assumed Mean of X
dy= Y - Assumed Mean of Y
N = Number of pairs of observations
Illustration-3:
Calculate Karl Pearson’s co-efficient of correlation from the following data.
X: 5, 10, 5, 11, 12, 4, 3, 2, 7, 6
Y: 1, 6, 2, 8, 5, 1, 4, 6, 5, 2
Solution:
X Y
dx
= X – A
Dy
= Y – A
dx
2
dy
2
dxdy
5
10
5
11
12
4
3
2
(7)
6
1
6
2
8
(5)
1
4
6
5
2
-2
3
-2
4
5
-3
-4
-5
0
1
-4
1
-3
3
0
4
-1
1
0
3
4
9
4
16
25
9
16
25
0
1
16
1
9
9
0
16
1
1
0
9
8
3
6
12
0
12
4
-5
0
3
∑ -5 ∑
∑ = 109 ∑ = 62 ∑
Correlation,
( )( )
√ √
=
√ √
=
= = 0.52
Merits
1. It gives an idea about the co-variation of the two series
2. It indicates the direction of relationship also
3. It provides a numerical measurement of co-efficient of correlation
4. It can be used for further algebraic treatment
5. It gives a single figure to explain the accurate degree of correlation
between two variables
Demerits
1. It assumes a linear relationship between the variables. But, in real
situations, it may not be so.
2. A high degree of correlation does not mean that a close relation exists
between variables.
3. Difficult to calculate.
4. It is unduly affected by extreme values.
PROBABLE ERROR
The quantity
( )
√
is known as the standard error of correlation co-
efficient. Usually, the correlation co-efficient is calculated from samples. For
different samples drawn from the same population, the co-efficient of correlation
may vary. But, the numerical value of such variation is expected to be less than
the probable error. It is a statistical measure which measures reliability and
dependability of the values of co-efficient of correlation. If probable error is
‘added to’ or ‘subtracted from’ the co-efficient of correlation, it would give two
such limits within which we can reasonably expect the value of co-efficient of
correlation to vary.
The probable error of the co-efficient of correlation can be obtained by
applying the formula:
Probable Error =
( )
√
6. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 6
If the value of r is less than the probable error, it is not at all significant. If
the value of r is more than six times of the probable error, it is significant. (If the
Probable Error is not much and if the value of r is 0.5 or more, it is generally
considered to be significant)
Uses
1. It is used to determine the limits within which the population correlation
co-efficient may be expected to lie.
2. It can be used to test if an observed value of sample correlation co-
efficient is significant of any correlation in population.
Spearman’s Rank Correlation
Karl Pearson’s correlation co-efficient is used to measure the correlation
between variables which are normally distributed. If population is not normal, or
the shape of the distribution is not known, Rank correlation is used. There are
many occasions whereby the value of certain variables cannot be measured in
quantitative form. For example, intelligence, beauty, character, morality, honesty,
etc. rank correlation is used to study association between such variables. It is a
method used to study the correlation between attributes. It was developed by
the British psychologist Charles Edward Spearman in 1904.
Cases
a) Ranks are not repeating
b) Repeated ranks/Tie in rank
Ranks are not repeating
STEPS:
1. Assign ranks to attributes
2. Compare the difference of ranks which is denoted by D
3. Calculate ƩD2
4. Apply the formula, and find correlation
R =
∑
Illustration-4:
Find out Spearman’s rank correlation co-efficient from the following data
X: 60, 34, 30, 50, 45, 41, 22, 43, 42, 66, 64, 46
Y: 75, 32, 34, 40, 45, 33, 12, 30, 36, 72, 41, 57
Solution:
X Y R1 R2 D D
2
60
34
40
50
45
41
22
43
42
66
64
46
75
32
34
40
45
33
12
30
36
72
41
57
3
11
10
4
6
9
12
7
8
1
2
5
1
10
8
6
4
9
12
11
7
2
5
3
2
1
2
-2
2
0
0
-4
1
-1
-3
2
4
1
4
4
4
0
0
16
1
1
9
4
∑
Rank Correlation, R =
∑
=
=
=
= 1- 0.167
= 0.833
Repeated ranks/Tie in rank
STEPS:
1. Assign ranks to attributes
2. Compare the difference of ranks which is denoted by D
3. Calculate ƩD2
4. Calculate m3
- m
5. Apply the formula, and find correlation
7. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 7
R =
⌊∑ ( )⌋
Illustration-5:
Find out Spearman’s rank correlation co-efficient from the following data
X: 68 64 75 50 64 80 75 40 55 64
Y: 62 58 68 45 81 60 68 48 50 70
Solution:
X Y R1 R2 D D
2
68
64
75
50
64
80
75
40
55
64
62
58
68
45
81
60
68
48
50
70
4
6
2.5
9
6
1
2.5
10
8
6
5
7
3.5
10
1
6
3.5
9
8
2
-1
-1
-1
-1
5
-5
-1
1
0
4
1
1
1
1
25
25
1
1
0
16
∑
m = 2, m3
– m = 23
– 2 = 8 – 2 = 6
m = 3, m3
– m = 33
– 2 = 27 – 3 = 24
m = 2, m3
– m = 23
– 2 = 8 – 2 = 6
Rank Correlation, R =
⌊ ( )⌋
=
=
= 1- 0.45
= 0.55
Merits
1. In this method, the sum of the differences between R1 and R2 is always
equal to zero. So it provides a check on the calculation.
2. It does not assume normality in the universe from which samples has
been drawn.
3. It is easy to understand and apply.
4. It is the way of studying correlation between qualitative data which
cannot be measured in quantitative terms.
Demerits
1. It cannot be measured in two-way frequency tables.
2. It can be conveniently used only when n is small.
3. Further algebraic treatment is not possible.
4. It is only approximate measure as the actual values are not used.
CO-EFFICIENT OF DETERMINATION
It is the square of co-efficient of correlation. It is more useful to measure
the percentage variation in the dependent variables in relation to the
independent variable.
Co-efficient of determination = r2
Or
=
The co-efficient of determination is a much useful and better measure of
interpreting the value of r. it states what percentage of variations in the
dependent variable is explained to be the dependent variable. If the value of r is
0.8, we cannot conclude that 80% of the value of the variation in the dependent
variable is due to the variation in the independent variable. The co-efficient of
determination in this case is r2
= 0.64 which implies that only 64% of variation in
the dependent variable has been explained by the independent variable and the
remaining 36% of variation is due to other factors.
8. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 8
MODULE-II
NON-PARAMETRIC TESTS
DEFINITION OF STATISTICS
The word statistics has been originated from the Latin word Status or the
Italian word Statista which means political state. According to Dr. S.P Gupta,
”Statistics is the science of collection, organization, presentation, analysis and
interpretation of numerical data”. Statistics can be divided into two branches:
1. Descriptive Statistics
2. Inferential Statistics
Descriptive Statistics deals with collection of data, its presentation in various
forms (tables, graphs, diagrams, etc.) and finding averages and other measures
which would describe the data. It refers to statistical techniques used to
summarize and describe a data set and the statistics (measures) used in such
summarizes.
Inferential Statistics deals with techniques used for analysis of data, making
the estimates and drawing conclusions from limited information taken on sample
basis and testing the reliability of estimates. It is the body of statistical techniques
that deals with the question “how reliable is the conclusions or estimates that we
derive from a set of data?”
PARAMETER AND STATISTIC
Parameter is a function of population values. It is a statistical measure
derived from the population. For example, arithmetic mean of a population.
Statistic is a function of sample values. It is a statistical measure derived
from the sample. For example, arithmetic mean of a sample.
Inferential statistics tries to predict the unknown parameter from the
known statistic.
TESTING OF HYPOTHESIS
Hypothesis is a statement subject to verification. It is an assumption made
about a population parameter. Lundberg defines hypothesis as a “tentative
generalization, the validity of which remains to be tested”. It is tentative, because
its veracity can be evaluated only after it has been tested empirically. It is stated
as an affirmative statement.
Hypothesis may be null hypothesis or alternative hypothesis and
directional hypothesis or non-directional hypothesis.
Null hypothesis is the original hypothesis. It states that there is no
significant difference between sample and population regarding a particular
matter under consideration. It is denoted by H0. For example, H0: There is no
significant mean difference in the mechanical aptitude between boys and girls
(µ1=µ2). Any hypothesis other than a null hypothesis is called alternative
hypothesis. It is denoted by H1. For example, H1: There is significant difference in
the mechanical aptitude between boys and girls (µ1≠µ2)
The statement ‘boys are better than girls in mechanical aptitude’ is a
directional hypothesis, as there is a clear indication of direction of change. But the
statement ‘boys and girls differ in mechanical aptitude’ is a non-directional
hypothesis, as there is no indication of direction of change.
PROCEDURE FOR TESTING HYPOTHESIS
Following are the various steps in the test of hypothesis.
1. Set-up hypotheses
Normally, the researcher has to set two types of hypotheses, viz; a null
hypothesis and an alternative hypothesis.
2. Set-up a suitable level of significance
The probability of rejecting a null hypothesis when it is true is known as
the level of significance. In other words, it is the probability of Type-I
error. Generally the level of significance is fixed at 5% or 1% (0.05 or 0.01).
Level of significance is denoted by α
3. Decide a test criterion.
The test criterion may be z-test, f-test, Ҳ2
-test, etc
4. Determine the degree of freedom
Degree of freedom is defined as the number of independent observations
which is obtained by subtracting the number of constraints from the total
number of observations. That is, degree of freedom = Total number of
observations – Number of constraints.
9. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 9
5. Calculation of test statistic
Test statistic can be calculated by using the formula, Difference/Standard
Error
6. Obtain table value
Table value is obtained by considering both the level of significance and
the degree of freedom.
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject H0
and accept H1. If the calculated value of test statistic is less than the table
value, we accept H0 and reject H1.
TYPE-I AND TYPE-II ERROR
While testing a hypothesis, the decision is to accept or reject a hypothesis.
Therefore, there are four possibilities of decisions:
1. Accepting a null hypothesis when it is true
2. Rejecting a null hypothesis when it is false
3. Rejecting a null hypothesis when it is true
4. Accepting a null hypothesis when it is false
The first and second cases are correct and the third and fourth cases are
errors. The third case is known as Type-I error and the fourth case is known as
Type-II error. That is the error of rejecting H0 when it is true is Type-I error, and
the error of accepting H0 when it is false is Type-II error. Type-II error is more
serious error.
The probability of Type-I error, that is rejecting a null hypothesis when it is
true is known as level of significance. As the probability of Type-I error decreases,
probability of Type-II error increases and vice versa.
REJECTION REGION AND ACCEPTANCE REGION
The entire area under a normal curve may be divided into two parts. They
are:
1. Rejection region, and
2. Acceptance region
Rejection Region
It is the area which corresponds to the predetermined level of
significance. If the computed value of the test statistic falls in the rejection region,
we reject the null hypothesis. Rejection region is also known as critical region. It is
denoted by α.
Acceptance Region
It is the area which corresponds to 1-α. If the computed value of the test
statistic falls in the acceptance region, we accept the null hypothesis
TWO-TAILED TEST AND ONE-TAILED TEST
Two-tailed Test
A two tailed test is one in which we reject the null hypothesis, if the
computed value of the test statistic is significantly greater than or lower than the
table value. In two-tailed test, the rejection region is represented by both tails,
that is left and right tails. If we are testing the hypothesis at 5% level of
significance, the size of the acceptance region is 0.95 and the size of the rejection
region is 0.05 on both sides together. So, if the computed value of test statistic
falls either in the left tail or in the right tail, the null hypothesis is rejected.
For example, if we want to test the null hypothesis that the average
height of people in the population is 156 cm. Then the rejection would be on both
sides, since the null hypothesis is rejected if the average height in the sample is
much more than 156 cm or much less than 156 cm.
One-tailed Tests
In one-tailed test, the rejection region is represented by one tail, which
may be either left tail or right tail. For example, if we want to test the null
hypothesis that average height of people in the population is more than 156 cm.
then the rejection area would be on the right tail only, since the null hypothesis is
rejected if the average height in the sample is much less than 156 cm.
Similarly, if we want to test the null hypothesis that average height of
people in the population is less than 156 cm. Then the rejection area would be on
the left tail only, since the null hypothesis is rejected if the average height in the
sample is much more than 156 cm.
10. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 10
TEST STATISTICS
The decision to accept or to reject a null hypothesis is made on the basis
of a statistic computed from the sample. Such a statistic is called test statistic. Test
statistic can be classified into two groups:
1. Parametric tests, and
2. Non parametric tests
Parametric Tests
The statistical tests based on the assumption that the population or
population parameter is normally distributed are called parametric tests. The
important parametric tests are:
a) z-test
b) f-test
c) t-test
Non Parametric Tests
It is a test which is not concerned with testing of parameters and does not
depend on the particular form of the distribution of the population. It can be
defined as a distribution free statistical test where assumptions are fewer than
those associated with parametric test. It is used when the researcher concludes
that a parametric test is not applicable.
Assumptions of Non-Parametric Test
1. The sample observations are independent
2. The variables are continuous
3. Sample drawn is a random sample
4. Observations are measured on ordinal scale
Merits of Non-Parametric Test
1. Simple and easy to apply
2. There is no assumption about the probability distribution of the
population
3. Non restriction regarding the size of sample
4. It can be used even if the sample is small
Types of Non Parametric Tests
1. Chi-square test (Ҳ2
-test)
2. Sign test
3. Signed rank test
4. Rank sum test
5. Runs test
CHI-SQUARE TEST (Ҳ2
-TEST)
It is a statistical test which explains the significance of difference between
a set of observed frequencies and a set of corresponding theoretical frequencies
under certain assumptions. It is a test which is not concerned with testing of
parameters and does not depend on the particular form of the distribution of the
population. It was developed by Prof: Karl Pearson in 1900.
Characteristics of Chi-Square Test
1. It is a non-parametric test
2. It is a distribution-free test
3. It is easy to evaluate chi-square test statistic
4. It analyses the difference between a set of observed frequencies and a set
of corresponding expected frequencies
Uses/Applications of Chi-Square Test
1. It is useful for the test of independence of attributes: Chi-square test can
be used to find out whether two attributes are associated or not.
2. It is useful for the test of goodness of fit: Chi-square test can be used to
ascertain how well the theoretical distribution fit the data.
3. It is useful for the testing of homogeneity: Test of homogeneity is
concerned with whether different samples come from the same
population.
4. It is useful for the testing given population variance: It helps to test
whether given population variable is acceptable on the basis of samples
drawn from that population.
CHI-SQUARE TEST FOR GOODNESS OF FIT
Chi-square test is used for testing hypothesis related to sample
proportions with respect to the corresponding population properties. Chi-square
11. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 11
test for goodness of fit determines how well they obtained sample proportions fit
the population proportions specified by the null hypothesis
Steps:
1. Set-up hypotheses
In test of goodness of fit, the hypotheses will be set as follows.
H0: There is goodness of fit between expected frequencies and observed
frequencies.
H1: There is no goodness of fit between expected frequencies and
observed frequencies.
2. Set-up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Decide a test criterion.
The test criterion will be chi-test
4. Determine the degree of freedom
Degree of freedom C-1, where C stands for the number of categories
5. Calculation of test statistic
Test statistic, Ҳ2
= [∑(O-E)2
]/E
Where, O is the observed frequencies, and E is the expected frequencies
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom.
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject H0
and accept H1. If the calculated value of test statistic is less than the table
value, we accept H0 and reject H1
CHI-SQUARE TEST OF INDEPENDENCE
Chi-square test is used for testing whether the two variables associated or
not.
Steps:
1. Set-up hypotheses
In test of independence, the hypotheses will be set as follows.
H0: The two attributes are independent.
H1: The two attributes are dependent.
2. Set-up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Determine the degree of freedom
Degree of freedom (R-1, C-1), where R stands for Number of rows, and C
stands for the number of columns.
4. Decide a test criterion.
The test criterion will be chi-test
5. Calculation of test statistic
Test statistic, Ҳ2
= [∑(O-E)2
]/E
Where, O is the observed frequencies, and E is the expected frequencies
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom.
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject H0
and accept H1. If the calculated value of test statistic is less than the table
value, we accept H0 and reject H1
Contingency Table:
A contingency table is a frequency table in which a sample from the
population is classified according to two or more attributes, which are divided into
two or more column. When there are only two divisions for each attributes, the
contingency table is known as 2X2 contingency table. The frequencies appearing in
the contingency table are known as cell frequencies.
For example, a 2X2 contingency table based on the two attributes
smoking and drinking is:
Smokers Not smokers Total
Drinkers 40 30 70
Not drinkers 4 24 28
12. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 12
Total 44 54 98
Calculation of Expected Frequencies:
Let a, b, c and d are the observed frequencies, and it is shown In the form
of a contingency table as follows:
Column 1 Column 2 Total
Row 1 A B f1
Row 2 C D f2
Total f3 f3 N
Then the expected frequencies are:
(f1 X f3)/N, (f1 X f4)/N, (f2 X f3)/N, (f2 X f4)/N, OR
(a + b) (a + c), (a + b) (b + d), (c + d) (a + c), (c + d) (b + d)
N N N N
SIGN TEST
Sign test is used to test whether the two populations are identical or not.
It is used in the situations, where t-test cannot be used. It is based on the direction
of the plus or minus signs of observations, and not on their numerical magnitudes.
The sign test may be:
1. One Sample Sign Test, and
2. Two Sample Sign Test
One Sample Sign Test
It is a very simple non-parametric test applicable when:
1) Sample is taken from a continuous population
2) P (Sample value ˂ Mean) = ½ and P (Sample value ˃ Mean) = ½
Steps:
1. Set-up hypotheses
In One Sample Sign Test, the hypotheses will be set as follows.
H0: P = ½
H1: P ≠ ½
2. Set-up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Decide a test criterion.
The test criterion will be One Sample Sign Test
4. Determine the degree of freedom
Degree of freedom is infinity.
5. Calculation of test statistic
Test statistic (p-P)/S.E
Where, p is the proportion of plus signs out of the total signs, P = ½, and
S.E = √ (PQ)/n, where Q = 1 – P
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom.
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject H0
and accept H1. If the calculated value of test statistic is less than the table
value, we accept H0 and reject H1
Illustration-1:
In a four round golf play scores of 11 professionals are 202, 210, 200, 203, 193,
203, 204, 195, 199, 202, and 201. Use one sample sign test at 5% level of
significance to test the null hypothesis that professional golfer’s average is 204.
Solution
X x-204
202 -
210 +
200 -
203 -
193 -
203 -
13. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 13
204 ….
195 -
199 -
202 -
201 -
H0: μ = 204
H1: μ ≠ 204
The level of significance is fixed at 5%.
The test criterion is One Sample Sign Test
Degree of freedom is infinity.
Test statistic = (p-P)/S.E
Where, p = 1/10, and P = ½
S.E =√ , that is, √ = √
Therefore, test statistic =
√
= -2.53
Table value at 5% level of significance and infinity degree of freedom is 1.96
As the calculated value of test statistic is more than (numerically) the table value,
we reject H0 and accept H1. That is, H1: μ ≠ 204
Two Sample Sign Test
Suppose X and Y are two variables and their n values are known. Then we
get n pair of values, first value of each pair being a value of X and the second is
that of Y. that is, if (x1, y1) is a pair, then X1 belongs to X and y1 belongs to Y.
In such cases, each pair can be replaced by + or – sign. If in a pair, first
value is greater than second value we put + sign. If first value is less than second
value we put - sign. If both are equal concerning value is discarded.
Steps:
1. Set-up hypotheses
In Two Sample Sign Test, the hypotheses will be set as follows.
H0: P = ½
H1: P ≠ ½
2. Set-up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Decide a test criterion.
The test criterion will be Two Sample Sign Test
4. Determine the degree of freedom
Degree of freedom is infinity.
5. Calculation of test statistic
Test statistic =
( )
√
( )
Where, p is the proportion of plus signs out of the total signs, and n is the
number of pairs compared.
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom.
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject H0
and accept H1. If the calculated value of test statistic is less than the table
value, we accept H0 and reject H1
Illustration-2:
The following are the numbers of tickets issued by two sales men on 11 days.
Sales man I: 7, 10, 14, 12, 6, 9, 11, 13, 7, 6, 10
Sales man II: 10, 13, 14, 11, 10, 7, 15, 11, 10, 9, 8
Use two Sample Sign Test at 1% level of significance to test the null hypothesis
that on the average the two sales men issue equal number of tickets
Solution
X Y Sign
7 10 -
10 13 -
14. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 14
14 14 …..
12 11 +
6 10 -
9 7 +
11 15 -
13 11 +
7 10 -
6 9 -
10 8 +
H0: μ = ½
H1: μ ≠ ½
The level of significance is fixed at 1%.
The test criterion is Two Sample Sign Test
Degree of freedom is infinity.
Test statistic =
( )
√
( )
=
( )
√
( )
= -0.63
Table value at 1% level of significance and infinity degree of freedom is 2.576
As the calculated value of test statistic is less than (numerically) the table value,
we accept H0, that is H0: μ = ½
SIGNED RANK TEST/WILCOXON MATCHED-PAIRS TEST
Signed Rank Test was originally proposed by Frank Wilcoxon in 1945. It is
a test used to evaluate the difference between the magnitude and signs of paired
observations. It can be used instead of T-test to produce a null hypothesis in cases
when the population does not conform to normal distribution.
CASE-I: When number of matched pairs (n) is less than 25
Steps:
1. Set up hypothesis
In Signed Rank Test, the hypothesis will be set as follows.
H0: There is no significant difference between two samples.
H1: There is significant difference between two samples.
2. Set up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Determine a test criterion
The test criterion will be Signed Rank Test.
4. Determine the degree of freedom
Degree of freedom is n.
5. Calculation of test statistic
Test statistic, T is lower of sum of ranks with sign.
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject
the H0 and accept H1. If the calculated value of test statistic is less than the
table value, we accept the H0 and reject H1.
Illustration-3:
Given below are 13 pairs of values showing the performance of two machines,
Test whether there is difference between the performances. Use Wilcoxon
Matched-Pairs Test.
Machine
A
73 43 47 53 58 47 52 58 38 61 56 56 34 55 65 75
Machine
B
51 41 43 41 47 32 24 58 43 53 52 57 44 57 40 68
Solution
Machine A Machine B
d
Difference
Rank with sign
15. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 15
73 51 22 13 …
43 41 2 2.5 …
47 43 4 4.5 …
53 41 12 11 …
58 47 11 10 …
47 32 15 12 …
52 24 28 15 …
58 58 0 … …
38 43 -5 … -6
61 53 8 8 …
56 52 4 4.5 …
56 57 -1 … -1
34 44 -10 … -9
55 57 -2 … -2.5
65 40 25 14 …
75 68 7 7 …
TOTAL 101.5 -18.5
H0: There is no significant difference between the performances of two machines.
H1: There is significant difference between the performances of two machines.
Set up a suitable level of significance
Here, the level of significance is fixed at 5%.
The test criterion is Signed Rank Test.
Degree of freedom is n. That is 15. As d=0 for the 8th
pair, it is not considered.
Test statistic, T=18.5 (Lower of 101.5 and 18.5).
Table value at 5% level of significance and 15 degree of freedom is 25
Since the calculated value of test statistic is less than the table value, we accept
the H0. That is, there is no significant difference between the performances of two
machines
CASE-II: When number of matched pairs (n) is greater than 25
Steps:
1. Set up hypothesis
In Signed Rank Test, the hypothesis will be set as follows.
H0: There is no significant difference between two samples.
H1: There is significant difference between two samples.
2. Set up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Determine a test criterion
The test criterion will be Signed Rank Test.
4. Determine the degree of freedom
Degree of freedom is infinity.
5. Calculation of test statistic
Test statistic, Z = (T-µ)/σ, where T is lower of sum of ranks with sign, µ = [n
(n+1)]/4, and σ = √ [n (n+1) (2n + 1)]/24
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject
the H0 and accept H1. If the calculated value of test statistic is less than the
table value, we accept the H0 and reject H1.
Illustarion-4:
The following are the weights in kilo grams, before and after of 26 babies who
stayed on a diet for some weeks.
Before: 7.0, 3.5, 2.1, 1.6, 7.5, 6.3, 7.0, 5.4, 7.7, 8.2, 6.8, 1.9, 1.3, 7.2, 7.8, 1.7, 2.4,
3.5, 4.5, 8.0, 1.5, 2.0, 5.8, 6.5, 3.5, 5.2
After: 7.9, 6.2, 9.0, 3.7, 3.5, 1.4 , 2.6, 3.2, 9.0, 5.4, 8.5, 4.4, 8.3, 9.0, 9.2, 3.2, 3.4
2.8, 3.4, 7.9, 3.5, 3.2, 6.2, 6.3, 3.0, 6.8
16. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 16
Solution
Before After
d
Difference
Rank with sign
7.0 7.9 -0.9 … -6
3.5 6.2 -2.7 … -20
2.1 9.0 -6.9 … -25
1.6 3.7 -2.1 … -17
7.5 3.5 +4.0 22 …
6.3 1.4 +4.9 24 …
7.0 2.6 +4.4 23 …
5.4 3.2 2.2 18 …
7.7 9.0 -1.3 … -10
8.2 5.4 +2.8 21 …
6.8 8.5 +1.7 … -14
1.9 4.4 -2.5 … -19
1.3 8.3 -7.0 … -26
7.2 9.0 -1.8 … -15
7.8 9.2 -1.4 … -11
1.7 3.2 -1.5 … -12
2.4 3.4 -1.0 … -7
3.5 2.8 +0.7 5 …
4.5 3.4 +1.1 8 …
8.0 7.9 +0.1 1 …
1.5 3.5 -2.0 … -16
2.0 3.2 -1.2 … -9
5.8 6.2 -0.4 … -3
6.5 6.3 +0.2 2 …
3.5 3.0 +0.5 4 …
5.2 6.8 -1.6 … -13
TOTAL 128 -223
H0: There is no significant difference between the weights of babies before and
after the diet.
H1: There is significant difference between the weights of babies before and after
the diet.
Here, the level of significance is fixed at 5%.
The test criterion is Signed Rank Test.
Degree of freedom is infinity.
Test statistic, Z= (T-µ)/σ
T = 128, that is, the lowest of 128 and 223
µ = [n (n+1)]/2, that is = [26 X (26 + 1)]/4, = 175.5
σ = √ [n (n+1) (2n + 1)]/24, that is = √ [26 (26+1) (2 X 26 + 1)]/24, = 39.37
Therefore Z= (128 – 175.5)/39.37, that is = -1.21
Table value at 5% level of significance and infinity degree of freedom is 1.96
Since the calculated value of test statistic is less than the table value, we accept
the H0. That is, there is no significant difference between the weights of babies
before and after the diet.
MANN–WHITNEY–WILCOXON U TEST
It is also called the Wilcoxon rank-sum test (WRS), or Wilcoxon–Mann–
Whitney test. It is a nonparametric test of the null hypothesis that two samples
come from the same population against an alternative hypothesis, especially that
a particular population tends to have larger values than the other. It can be
applied on unknown distributions contrary to t-test which has to be applied only
on normal distributions, and it is nearly as efficient as the t-test on normal
distributions.
Let x1, x2, …..,xn be the values of X variable and y1, y2, ….., yn be the values
of Y variable. Let the values of X form a sample independent of the sample formed
by values of Y. we want to test whether the two samples have come from two
identical populations. Let the probability function of X be f1 (x) and that of y be f2
(y).
17. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 17
Steps:
1. Set up hypothesis
In Mann–Whitney–Wilcoxon U Test, the hypothesis will be set as follows:
H0: The populations are identical, that is f1 (x) = f2 (y)
H1: The populations are not identical, that is f1 (x) ≠ f2 (y).
2. Set up a suitable level of significance
Generally, the level of significance is fixed at 5% or 1%.
3. Determine a test criterion
The test criterion will be Mann–Whitney–Wilcoxon U Test.
4. Determine the degree of freedom
Degree of freedom is infinity.
5. Calculation of test statistic
Test statistic, t =(µ - U)/S.E, where µ = (n1 X n2)/2, U = n1 X n2 + [ n1 X (n1 +
1)/2] – R1, and S.E = √*n1 X n2 (n1 + n2 + 1)]/12
6. Obtain table value
Table value is obtained by considering both level of significance and
degree of freedom
7. Making decision
The decision may be either to accept or to reject the null hypothesis. If the
calculated value of test statistic is more than the table value, we reject
the H0 and accept H1. If the calculated value of test statistic is less than the
table value, we accept the H0 and reject H1.
Illustration-5:
There are two samples. First contains the observations54, 39, 70, 58, 47, 40, 74,
49, 74, 75, 61, and 79. The second contains 45, 41, 62, 53, 33, 45, 71, 42, 68, 73, 54,
and 73. Apply Mann–Whitney–Wilcoxon U Test at 5% level of significance that
they come from populations with the same mean?
Solution
Values
(Ascending order)
Rank
Sample
I or II
33 1 II
39 2 I
40 3 I
41 4 II
42 5 II
45 6.5 II
45 6.5 II
47 8 I
49 9 I
53 10 II
54 11.5 II
54 11.5 I
58 13 I
61 14 I
62 15 II
68 16 II
70 17 I
71 18 II
73 19.5 II
73 19.5 II
74 21.5 I
74 21.5 I
75 23 I
79 24 I
H0: The populations are identical, that is f1 (x) = f2 (y)
H1: The populations are not identical, that is f1 (x) ≠ f2 (y).
The level of significance is fixed at 5%.
The test criterion is Mann–Whitney–Wilcoxon U Test.
Degree of freedom is infinity.
Test statistic, t = (µ - U)/S.E
µ = (n1 X n2)/2, that is, = (12 X 12)/2 = 72
R1 = Sum of ranks assigned to the values in sample I, that is = 167.5
18. Psychological Statistics-III
Irshadiya College of Commerce and Social Sciences, Feroke Page 18
U = n1 X n2 + [ n1 X (n1 + 1)/2] – R1, that is, 12 X 12 + [ 12 X (12 + 1)/2] – 167.5, =
54.5
S.E = √*n1 X n2 (n1 + n2 + 1)+/12, that is, √*12 X 12 (12 + 12 + 1)+/12, = 17.32
Therefore, t = (72–54.5)/17.32 = 1.01
Table value at 5% level of significance and degree of freedom infinity is 1.96
As the calculated value of test statistic is less than the table value, we accept the
H0 and reject H1, that is, the populations are identical, that is f1 (x) = f2 (y)
********************