The material is consolidated from different sources on the basic concepts of Statistics which could be used for the Visualization an Prediction requirements of Analytics.
I deeply acknowledge the sources which helped me consolidate the material for my students.
Introduction to Statistics - Basic concepts
- How to be a good doctor - A step in Health promotion
- By Ibrahim A. Abdelhaleem - Zagazig Medical Research Society (ZMRS)
For a detailed explanation Watch the Youtube video:
https://youtu.be/YK0GPKuYVfU
Classification of Data, methods- geographical classification, Chronological Classification, Qualitative and Quantitative classification, discrete and continuous variable, grouped frequency distribution, inclusive , exclusive series, cumulative frequency distribution
Introduction to Statistics - Basic concepts
- How to be a good doctor - A step in Health promotion
- By Ibrahim A. Abdelhaleem - Zagazig Medical Research Society (ZMRS)
For a detailed explanation Watch the Youtube video:
https://youtu.be/YK0GPKuYVfU
Classification of Data, methods- geographical classification, Chronological Classification, Qualitative and Quantitative classification, discrete and continuous variable, grouped frequency distribution, inclusive , exclusive series, cumulative frequency distribution
Lecture on Introduction to Descriptive Statistics - Part 1 and Part 2. These slides were presented during a lecture at the Colombo Institute of Research and Psychology.
hello, friends. i'm humaira jahan. and this is my presentation on statistics. you can find a overall concept of statistics in this presentation. and i hope it will help you enough to know about statistics.
presentation of data
Tabulated data can be easily understand and interpreted.
Graphical forms makes it possible to easily draw visual impression of data.
It makes comparisons easily.
This kind of method create an imprint on mind for a long period of time.
Understanding data type is an important concept in statistics, when you are designing an experiment, you want to know what type of data you are dealing with, that will decide what type of statistical analysis, visualizations and prediction algorithms could be used.
#data #data types #ai #machine learning #statistics #data science #data analytics #artificial intelligence
Presentation is made by the student of M.phil Jameel Ahmed Qureshi Faculty of Education Elsa Kazi campus Hyderabad UoS Jamshoron, This presentation is an assignment assign by the Dr. Mumtaz Khwaja
Lecture on Introduction to Descriptive Statistics - Part 1 and Part 2. These slides were presented during a lecture at the Colombo Institute of Research and Psychology.
hello, friends. i'm humaira jahan. and this is my presentation on statistics. you can find a overall concept of statistics in this presentation. and i hope it will help you enough to know about statistics.
presentation of data
Tabulated data can be easily understand and interpreted.
Graphical forms makes it possible to easily draw visual impression of data.
It makes comparisons easily.
This kind of method create an imprint on mind for a long period of time.
Understanding data type is an important concept in statistics, when you are designing an experiment, you want to know what type of data you are dealing with, that will decide what type of statistical analysis, visualizations and prediction algorithms could be used.
#data #data types #ai #machine learning #statistics #data science #data analytics #artificial intelligence
Presentation is made by the student of M.phil Jameel Ahmed Qureshi Faculty of Education Elsa Kazi campus Hyderabad UoS Jamshoron, This presentation is an assignment assign by the Dr. Mumtaz Khwaja
This presentation on Introduction to Statistics helps Engineering students to review the fundamental topics of statistics. It is according tl syllabus of Institute of Engineering (IOE) but is similar to that of almost all the engineering colleges.
This presentation covers statistics, its importance, its applications, branches of statistics, basic concepts used in statistics, data sampling, types of sampling,types of data and collection of data.
1. Introduction to statistics in curriculum and Instruction
1 The definition of statistics and other related terms
1.2 Descriptive statistics
3 Inferential statistics
1.4 Function and significance of statistics in education
5 Types and levels of measurement scale
2. Introduction to SPSS Software
3. Frequency Distribution
4. Normal Curve and Standard Score
5. Confidence Interval for the Mean, Proportions, and Variances
6. Hypothesis Testing with One and two Sample
7. Two-way Analysis of Variance
8. Correlation and Simple Linear Regression
9. CHI-SQUARE
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
1. STATISTICS FOR DATA ANALYTICS
Dr.S.Saudia,
Assistant Professor,
CITE, M.S.University
2. Objectives of the Chapter.
• To understand the different Descriptive Statistics concepts for
Descriptive Analytics and later analytics stages.
3. Chapter contents.
• Sampling Techniques: Simple Random Sampling, Convenience, Systematic,
Cluster and Stratified
•Organization of Data: Data classification, Tabulation, Frequency and
Graphic representation
• Measures of central value: Arithmetic mean, Geometric mean, Harmonic
mean, Mode, Median, Quartiles, Deciles, Percentile
• Measures of variation: Range, IQR, Quartile deviation, Mean deviation,
standard deviation, coefficient variance, Skewness, Moments & Kurtosis.
4. • Population: Set of persons or objects which are investigated for statistical study and research.
Eg: the students in M.S.U. , Books in a library (finite population)
Bulbs produced by factory this year (can be used as the population for analysis next year also)
(Hypothetical population)
A parameter is an unknown numerical summary of a population. Eg: The proportion of
diabetic patients feeling fatigue atleast once in a week.
• Sample: Part of the population from which information is collected. Always not all the information is
collected from the entire population for analysis.
A statistic value is a known numerical summary of a sample of the population. Eg: The
proportion of diabetic patients feeling fatigue often.
Statistics
Statistics is a methodology for collecting, analyzing, interpreting and drawing conclusions from numerical or
categorical data [1]. It provides methods for i) summarizing/ organizing data
ii) exploring data (Is the new grain productive?)
iii)Making predictions from data (What will be the employment rate
next year?)
Basic concepts of Statistics [1]
5. • Descriptive Statistics
Types of Statistics[1],[4]
• The Statistical methods used for summarizing and describing information in data.
• It includes construction of graphs, charts and tables, calculation of various descriptive measures:
average, measures of variation and percentiles.
Eg: For an event of tossing a dice, descriptive statistics can be summarizing the frequencies of outcomes
as:
• The Statistical methods used for drawing and the reliability of conclusions about population based on
the information obtained from the sample population.
• It includes methods for point estimation, interval estimation, hypothesis testing using probability theory.
Eg: For an event of tossing a dice, inferential statistics is verifying whether the dice is fair or not.
Inferential statistics is done on the information obtained from Descriptive Statistics. An inference is made on
the population based on information obtained from the sample.
• Inferential Statistics
Descriptive Statistics [1]
6. A parameter is an unknown numerical summary of a population. Eg: The proportion of diabetic
patients going fatigue in a week.
A statistic is a known numerical summary of a population. Eg: The proportion of diabetic patients
going fatigue in a week which is calculated from a sample.
The statistical values are used to find the unknown inferential parameters about the population.
Objective of Statistics methods is to make inferences about the population from an analysis of information
contained in sample data.
• Parameter and Statistics
Parameter and Statistic[2]
7. Nominal Scale: the categories of the qualitative variables are in an unordered scale.
There is no order for these categories or only naming of the categories happen in this scale.
It cannot be said that one category is greater/ lesser than the other.
Eg: Gender, Marital Status.
The difference between the categories need be not the same or it is unknown.
These categories can be represented as pie chart or bar chart.
Can you calculate the mean/ median of a nominal variable? Why?
Measures of central tendency for nominal variables is mode.
Nominal Scale of Gender, Mode of Transport [3]
Variable is a characteristic that varies from one person to person/ object to object of a population. Eg: height,
weight.
The variables can be Quantitative or Qualitative variables.
Qualitative Variables: These variables will have categorical values.
These variables are measured on a nominal scale or ordinal scale or interval scale or ratio scale.
• Variable
8. Ordinal Scale: the categories of the qualitative variables are on an ordered scale. Low to High
ordering/ ranking is possible for such categories.
It can be said that one category is more than other category but the difference between the categories is unknown.
Eg: Education rating, Temperature, Strength of opinion polls, customer satisfaction (from very satisfied, more satisfied,
satisfied, dissatisfied etc.), Pain Intensity etc.
A first, second, third kind of ranking
and there is no absolute zero.
The median and mode of the categories can be calculated.
No arithmetics can be done.
Ordinal Scale of temperature [3]
Movie Ratings [4]
9. Quantitative Variables: These variables have numeric values.
When the numeric values have discrete finite values they are called discrete quantitative
variables. Eg: Children in a family.
When the numeric values can be expressed more accurately they are called continuous
quantitative variables. Eg: Height, Weight.
These variables are measured on an interval scale or ratio scale.
Interval Scale: the values of the quantitative variables are on an ordered scale and the difference
between the interval values is known and is equal.
Value zero in an interval scale is arbitrary which means zero is also a measurement and not that
the measurement value is zero.
Eg: A temperature at 0 degree celsius does not mean that there is no temperature, Time.
The difference between the points on the
scale is
measurable and equal.
So mean,
median and mode of the categories can be
calculated.
Also addition and subtraction can be done.
Interval Scale of
temperature in
degree celsius [6]
10. Ratio Scale: the values of the quantitative variables are on an ordered scale and the difference
between the interval values is known and is equal.
Value zero in an interval scale is absolute and occurs naturally as a measurement.
Eg: A temperature at 0 Kelvin does not mean there is no temperature, it means that the temperature is zero., height, weight
etc.
All statistical measurements and all arithmetics can be
done.
Ratio Scale of
height [7]
Decision Tree To decide on the levels/ scale of measurement [4]
12. Order of Scales of Measurement [6]
In the highest level/ scale of a data representation which is the numeric ratio level, all the measurements can be
done on the data-set.
13. The value of a variable for one or more people or things yield data.
• Data
The collection of all observations for a particular variables is called a dataset/ data matrix. The data
sets are recorded for the samples of the population of interest.
All the values of the data set is presented as a data matrix as shown below.
The values of the different variables are arranged in different columns.
Each row correspond to the values of variables for a distinct sampling unit.
xij is the value of the jth variable and the ith sample. i=1,2…n, j=1,2,…k
• Data Set/ Data Matrix
Data Matrix [1]
14. 1. Sampling Techniques [9]
Part of the population from which information is collected is called a sample.
The entire population cannot be studied and so is the sample selected for study.
There are different techniques of sampling. It depends mainly upon on the resources available like
time and money needed to conduct the study and also the nature of population.
The sampling techniques should select an unbiased, representative sample. Otherwise the sample shall produce
sampling error.
Eg: If the study is on the African and European population, the sample should have
representatives of Africans and Europeans
equal number of Europeans and Africans to be unbiased.
Factors on which Sampling Techniques depend upon. [7], [8]
DEPENDS
UPON
15. • Simple Random Sampling
This sampling technique involves the selection of random objects based on some random numbers as
the sample of a population. The method produces an unbiased and representative sample. Every object of the
sample is equally likely to be a member of the sample.
It is theoretically the most ideal method of sampling.
Simple Random Sampling [8], [9]
Simple random sampling can be quite expensive when the population corresponds to people across
the world. It is good for samples from manufacturing industry.
It is better sampling technique when sampling frame (list of all objects/ people present in the
population of interest) is small.
16. • Convenience Sampling
This sampling technique is a convenient sampling technique, samples are collected from nearest
neighborhood minding the time and cost limits.
Eg: The samples are from a shopping mall or from someone’s friends/ relatives/ people passing by.
The latest items produced from a manufacturing industry is selected to be a sample.
It produces very much biased (sometimes self selection bias- when people who are interested in the problem
participate in the survey) and un representative samples.
Convenience Sampling [10]
17. • Systematic Sampling
This sampling technique is a systematic sampling technique where the first object of the sample is
selected based on a random number and later objects of the samples are selected after a fixed number apart.
It is easier than Simple Random Sampling to produce a random approximation of the sample.
However if there is a pattern in the data sample, then a certain type of object getting selected more is common in
Systematic Sampling.
Systematic Sampling [11]
18. • Cluster Sampling
The population is first divided into clusters and then the clusters are chosen in random. All the objects
of the cluster are included in the sample.
Eg: Departments of a business.
It is more convenient than Simple Random Sampling.
If the clusters chosen are not more different from each other, the the so selected sample can be
biased or un representing the population.
Cluster Sampling [13]
19. • Stratified Sampling
The population is first divided into well defined strata and then the objects from each strata are
chosen in random for being the samples. No. of objects selected from each strata correspond to the size of the
strata.
Eg: If the population corresponds to the whole of the population, then the different nationals, other
characteristics like: age, occupation, culture etc. correspond to a strata.
Stratified Sampling [13]
21. 2. Organization of Data
The data collected need to be organized. The data can be organized by describing them as tables and
graphs.
From the table or the graph, the range or the spread (maximum –minimum value) of the data can be
obtained. Also the count of each value, the maximum value etc. can also be obtained.
Qualitative variable:
The number of observations of a particular qualitative variable is the frequency/count of that variable.
The frequency of the variable data can be organized as
1. the frequency distribution-a table listing all the classes and its frequencies.
2. percentage of a class-frequency of the class to the total number of observations x100%
3. relative frequency-frequency of the class in decimal
4. relative frequency distribution-a table listing all the classes and their relative frequencies. Sample
size decides the credibility of relative frequency. The relative
frequencies add up to 1.
Data represented as a Table, Data represented as a bar graph [15].
22. 5. cumulative frequency-sum of frequencies up to a particular class. It gives the frequency
above or below a reference level.
6. relative cumulative frequency-sum of relative frequencies up to a particular class.
7. pie chart: The qualitative data is graphically represented as a pie-chart. It is a disc divided
into a number of pieces depending upon the frequencies of the classes. The
angle for a class is obtained by dividing the relative frequency by 360 degree.
Nominal data are represented by pie-chart .
8. bar graph: A bar graph is also a graphical representation where the classes are displayed
on the horizontal/ vertical axis and the frequencies are displayed on the
vertical/ horizontal axis. Accordingly they are called horizontal bar graph and
vertical bar graph respectively.
Ordinal data are represented as bar graph.
23. Example 3.1. Let the blood types of 40 persons are as follows:
O O A B A O A A A O B O B O O A O O A A A A AB A B A A O O A
O O A A A O A O O AB
Blood
group
Frequency Relative
Frequency
Cumulative
Frequency
O 16 .4 16
A 18 .45 34
B 4 .1 38
AB 2 .05 40
Total 40 1
Table 2 Relative and Cumulative Frequency [1]
Cumulative frequency calculation is more significant for ordinal and quantitative data.
Exercise: Write the R code for finding the frequency, relative frequency and cumulative
frequency of the blood group data
24. Pie Chart and Bar graph [1]
Exercise: Write the R code to plot the pie-chart of the blood group data
25. Quantitative variable:
The frequency of the quantitative variable data if they are less in number is calculated as for qualitative variables.
If the number of the data is quite large, the data are grouped into classes before calculating the frequency
distribution.
Generally 5-15 class intervals are chosen. Percent, cumulative Frequency, Relative Frequency etc. holds good even
for the quantitative data like the qualitative data.
1. histogram distribution- histogram for a grouped data displays the frequency or the relative
frequencies of each class interval. It is a bar graph of the grouped data.
26.
27. Exercise: Write the R code for finding the frequency, relative frequency and cumulative frequency
of the data and plot the histogram of the frequencies
28. 2. O give: A cumulative frequency can be visualized using a curve called an Ogive. Ogives can be
plotted against the upper or the lower limits of their class intervals and accordingly they are called less than
Ogive or greater than Ogive.
For the given distribution the cumulative frequency curve/ O give is shown below.
Frequency Table , Frequency distribution and Ogive [14] for the population whose age is recorded [14]
29. Points to Remember:
The frequency distribution of the population is called a population distribution and that of a sample is called a
sample distribution.
The sample distribution is a blurred image of the population.
As the sample size increases, the sample relative frequency is very close to that of the population and image
becomes clearer.
Also as the sample size increases, the histogram curve of the frequency distribution becomes more smooth/
continuous as can be seen below.
Histograms in order for sample sizes 100, 2000 and the whole population [1]
30. Points to Remember:
A summary of the population can be made by looking at the shape of the distribution curve.
As shown below for the bell and U shaped distributions, the populations are very clearly different.
U shaped and Bell shaped Histograms of two different populations [1]
31. Points to Remember:
Bell and U shaped distributions of populations are symmetric.
The distributions can also be skewed to one direction, left skewed or right skewed as shown below. These are
asymmetrical distributions.
Distributions skewed to the right and Distribution skewed to the left [1]
32. 3. Measures of central values [1]
That single value/measure which can be representative (a typical value) of the whole population is called the
measure of central tendency. It is the middle of a population.
The three measures of central tendencies are:
1. Mean
2. Median
3. Mode
Mean and median are measures of central tendencies for quantitative data only whereas mode is a
measure of central tendency for qualitative data.
For a symmetric population (normal distribution), these values are close to each other. Also if the population is
almost similar, these values will be same.
For an unsymmetrical or skewed population, these values are different from each other as shown in figure above.
Measures of Central Tendency of symmetrical and unsymmetrical populations[1]
33. Measures of Central Tendencies :
Mode (Common/ Popular Qualitative Variable)
Mode is used to find that value which occurs more often.
Eg: 1. Which product is sold more?
The mode gives idea about the popular product, a popular size/color of a cloth/any item. This
will help the retailer to stock that product more.
2. Which country is prone to CANCER?
Mode gives idea about the country with more CANCER patients. This will help the government
to take better precautionary measures.
Mode of a qualitative or discrete quantitative variable is the value of the variable with highest frequency.
If the greatest frequency is 1 then the variable has no mode.
If the greatest frequency is 2 or greater than 2, then the mode is equal to the greatest frequency.
The mode can be easily determined from the frequency distribution table or graph.
The mode in the frequency distribution table 6 [1] is A. This means that A is the most common blood group.
34. When the data is large, continuous and divided into classes, the mode is a mode class which is the class interval with
the highest frequency.
The mode class in the frequency distribution table 7 [1] is 0.065-0.085.
35. Median (Mid value)
Median of a quantitative variable is that central value which divides the ordered values of the variable set into a set
less than the median and a set greater than the median. The measure demands a data set which can be ordered.
When n is the number of items in the ordered data set,
If n is odd, median is the middle value in the ordered data set.
If n is even, median is either of middle values in the ordered data set.
From this measure an idea is obtained about the total number of values less than the central measure and the total
number of values above the central measure can be found. Both numbers are equal.
Measures of Central Tendency : Median[16]
1
or
2
1
2
n
Median floor
n
Median ceil
+
=
+
=
36. The measure of median is not affected by the outlier values (extreme values) in the data set.
Consider the data set below which are prices of some cottages in an area for sale.
The median ($137,500) from this data set is a measure from among the normal values ($125,000, $127,000, $135,000,
$140,000, $148,000, $150,000). in the data set and it not get affected by the extreme values ($110,000 and
$380,000).
Thus,
• Median is the best measure for asymmetrical data / for a skewed distribution or when the distribution is not
normal.
• It is not affected by all the values in the dataset and so is reliable when there are outliers.
• Good for data in ratio and interval scale. =
Data Set – Prices of cottages , Median calculation [15]
37.
38. Mean/ Arithmetic Mean
Mean of a quantitative variable is that common central value is the sum of all the observations of the variable divided
by the total number of observations.
When n is the number of items in the data set, the mean , is
_
x
39. The mean value calculated shall be influenced by the extreme values or outliers.
The mean ($164,375) from this data set is a measure from among all the values ($125,000, $127,000, $135,000,
$140,000, $148,000, $150,000, $110,000 and $380,000) in the data set. The value, $380,000 is observed to be double
the greatest value in the dataset.
Thus,
• Mean is the best measure for symmetrical data / for a normal distribution.
• It is affected by all the values in the dataset and so is reliable when there are no outliers.
• Good for data in ratio and interval scale.
Mean should have a logical connection to the expected value.
Data Set – Prices of cottages , Mean calculation [15]
40. Exercise: Write the R code for finding the mean, median and mode of the
data set:
41. Geometric Mean [19]-[20]
Geometric Mean of a quantitative variable is that value which is the nth root of the product of all the n observations.
The geometric mean is the average of a set of products.
It is commonly used
when working with percentages, which are derived from values,
For values over a time period like rates
But the standard arithmetic mean works with the values themselves.
Eg: Average returns from a stock market
to determine the performance results of an investment or portfolio.
Average rate under compound interest
Average Depreciation of machines/ shrinkage factors/ growth rate/ change factors
When n is the number of items in the data set,
or
where
Exercise
42. Why Geometric Mean over Arithmetic Mean [21]
Eg: To find the average interest rate. Consider that an amount of $ 100 is invested. If in the first year, an interest
percent of 10% is drawn and in the second year an interest of 20% and in the third year a 39%, then after three years,
the amount drawn shall be
=100(1+r1)(1+r2)(1+r3)
=100(1+.10)(1+.20)(1+.30)= $ 171.6
If it is arithmetic mean to find the average interest rate, ,
Then Arithmetic mean=10+20+30/3=20%
The amount drawn after using arithmetic mean rate of interest is 100(1+.2) (1+.2) (1+.2)=$ 172.8
The amount is more than the actual amount.
So another mean rate, r is to determined which must give the correct amount drawn at the end.
i.e, 100x(1+r) 3= 100(1+r1)(1+r2)(1+r3)
(1+r) 3= (1+r1)(1+r2)(1+r3)
Taking the cube root on either side,
Substituting values of r we get,
The amount drawn using this rate of interest is 100(1+..197) (1+..197) (1+..197)=$ 171.5
The amount is more similar to the actual amount.
So this rate, the geometric mean (nth root of the product of n items) is better in such average calculations.
( )( )( )3 1 .1 1 .2 1 .3 1r = + + + −
.197r =
43. Exercise:
Write the R code for finding the mean compound interest if the interests for the first
five years are 10%,20%, 30%, 40%, 50%.
Weighted Geometric Mean [22]
If have then the geometric mean is1, 2, 3, 4,.... nx x x x x 1, 2, 3, 4,.... nf f f f f
( )31 2
1 2 3
1
, , ... n
n
f ff f N
MG x x x x=
Where N is the count of all frequencies.
44. Harmonic Mean [21-[22]
Harmonic Mean of a quantitative variable is reciprocal of the arithmetic mean of the reciprocal of observations.
Consider a situation below where caps are ordered in bulk under three price categories.
If the average price for a cap is to be fixed, the arithmetic mean= 12+16+15/3 will not work out.
This is because the number of xi values ($12,$16, $15) is not equal to 3. It is
240/12 for the 1st cap
160/16 for the 2nd cap
300/15 for the 3rd cap
• Also the sum o f observations is not equal to $12+$16+$15. It is
12x240/12 for the 1st cap + 16x160/12 for the 2nd cap + 15x300/15 for the 3rd cap
So the arithmetic mean is =12x240/12 + 16x160/16 + 15x300/15 = 240 + 160 + 300 = W1+W2+W3
240/12 +160/16+ 300/15 240/12 +160/16+ 300/15 W1/X1+ W2/X2+ W3/X3
0 0
1
1 1 1M n n
i ii i
n
H
n x x= =
= =
∑ ∑
46. Quartiles are such measurements which divide the total number of observations into 4 equal parts. There are three
quartiles: First Quartile (or Lower Quartile):Q1
Second Quartile (Middle Quartile):Q2 / Median
Third Quartile (or Upper Quartile): Q3
The number of observations (N/4) smaller than Q1 is same as the number lying between Q1 and Q2, or between Q2 and
Q3, or larger than Q3.
For continuous observations, one quarter of the observations are smaller than Q1, two-quarters are smaller than Q2 and
three quarters are smaller than Q3.
So Q1, Q2 and Q3 are the values corresponding to cumulative frequencies, n/4, 2n/4, 3n/4 respectively for a grouped data
set.
Partition Values[1], [24]
The total number of observations can be divided into fixed number of parts, say 4 or 10 or 100 by values called
partition values/ quantiles.
Some of the quantiles are :
Median
Quartiles
Deciles
Percentiles
Quartiles
Quartiles [24]
47. Percentile
Percentiles are such measurements of the variable which divide the total number of observations into 100 equal parts.
The first percentile, P1 is that value of the variable which divides the bottom1% values from the top 99% values.
The second percentile, P2 is that value of the variable which divides the bottom2% values from the top 98% values.
The median is thus the 50th percentile, P50.
Percentile [25]
48. Deciles
Deciles are such measurements of the variable which divide the total number of observations into 10 equal parts.
There are 9 deciles, D1,D2, D3, D4, D5, D6, D7,D8, D9
The first decile, D1 is that value of the variable which divides the bottom10% values from the top 90% values. It is also the
10th percentile, P1.
Similarly, the second decile, D1 is the 20th percentile, P20 and so on.
Decile [26]
49. Five number Summary and Box Plot
Five number summary of a variable consists of minimum, maximum and three quartiles written in the increasing order.
They provide information on center and variation of variable.
Box plot is based on the five number summary and it gives a graphical display of the center and the variations.
Box plots can be in two types: 1. Box plots and 2. Modified Boxplot.
Outliers are marked in Modified Box plot and not in Boxplot.
Steps for drawing Boxplot.
50.
51. Exercise
Find the popular color using the relevant central measure.
Understand the significance of that central measure.
Data set [17]
52. Exercise
Find the measure of central value of this elephant population.
Write a statement about the population using that measur of central tendency. .
Data set [18]
53. Exercise
Find the measure of central value of these dogs’ height .
Comment on the variation of each heights from the cental measure calculated.
Data set [28]
55. 4. Measures of variation [1]
In addition to measuring the central values of a dataset, it also important to understand the variation of
each value from the central measure for any analytics. Such measurements are called Measures of Variation.
They are: 1. Range
2. Interquartile range
3. Standard Deviation
Range: The sample range of a variable is the difference between the maximum and minimum values
of the variable in the dataset.
Range= Max-Min
Range determined for a dataset cannot decrease but can increase when more values are added to the dataset.
56. Interquartile range : The sample interquartile range, IQR of a variable is the difference between the
first and third quartiles of that variable.
IQR=Q3-Q1
It is otherwise the range of the middle part of the dataset.
57. Standard Deviation [1]: It is a most commonly used as a measure of variability.
Standard Deviation, Sx can defined as the root of the average of the square of the absolute
deviations of observations from the mean of the variable.
Here mean/ average is used as the standard. The value is always positive.
The mean of all squared deviations between observations and mean of the observation is called Sample Variance,
Sx
2.
2
1
n
i
i
x
x x
S
n
σ
−
=
−
= =
∑
2
2 2 1
n
i
i
x
x x
S
n
σ
−
=
−
= =
∑
58. For normal distribution/ symmetric bell shaped population, it is experimentally determined that:
68% of the values lie within
95% of the values lie within
99.7% of the values lie within
Standard deviation fluctuates less when compared other measures of dispersion when moving from sample to
sample.
x σ
−
±
2x σ
−
±
3x σ
−
±
Normal Distribution with mean =0 [27]
59. Mean Deviation/ Mean Absolute Deviation [22]:
Mean Deviation can defined as the average of the absolute deviations of observations from
the mean/ any other specified value of the variable.
Generally, Mean deviation is calculated about Mean.
Mean Deviation about median is the least.
Exercise
Calculate the Mean Deviation of the following data about the median: 8,15,53,49,19,62,7,15,95,77
1
1
A
n
i
i
Mean Deviation about x A
n =
= −∑
Mean Deviation [28]
60. Moments : Moments about any arbitrary constant A are defined as
The four Moments about a zero are called Raw Moments
and the moments about mean are called Central Moments.
Raw Moments (about zero) Central Moments (about mean)
( )
( )
( )
' st
1
2' nd
2
3' rd
3
1
is the 1 moment
1
is the 2 moment
1
is the 3 moment
x A
n
x A
n
x A
n
µ
µ
µ
= −
= −
= −
∑
∑
∑
( )
( )
( )
' st
1
2' nd
2
3' rd
3
1
is the 1 moment about 0
1
is the 2 moment about 0
1
is the 3 moment about 0
x
n
x
n
x
n
µ
µ
µ
=
=
=
∑
∑
∑
st
1
2
nd
2
3
rd
3
4
th
4
1
is the 1 moment about
1
is the 2 moment about
1
is the 3 moment about
1
is the 4 moment about
x x x
n
x x x
n
x x x
n
x x x
n
µ
µ
µ
µ
− −
− −
− −
− −
= −
= −
= −
= −
∑
∑
∑
∑
61. Moments are used to describe the basic peculiarities of the data from its frequency distribution like
: measure of the central tendency is given by the first raw moment
measure of dispersion is given by the 2nd moment about mean
symmetry/ skewness of the curve is given by 3rd moment about mean
kurtosis is given by 4th moment about mean.
Moments [29]
62. Skewness [31] : Skewness mentions extent of asymmetry of a dataset. It also speaks about the direction
of variation of the dataset from the mean.
It can be Negative Skewness or Positive Skewness depending upon the if the distribution is skewed to the right or the left
of the dataset’s mean respectively.
The skewness will give knowledge as to how much is the dataset greater than or is less than the mean.
From the above figure, note the positions of Mean, Median and Mode. The difference between the Mean and Mode can
give the measure and direction of Skewness.
Normal distribution and Skewed distributions [32]
63. Measure of Skewness: To find extend of asymmetry and to give the direction (positive or negative)
Pearson’s first Measure:
Skewness is positive when Mean is larger than Median and Mode and Vice-versa.
Pearson’s second Measure: When mode is ill defined,
Bowley’s Measure:
Where are the three quartiles. is the median. For a positively skewed distribution, Q3 will be away
from Q2 and Q1 and vice-versa for a negatively skewed distribution.
The Bowley’s formula is used to calculate skewness when the dataset is a grouped dataset similar to the one mentioned
in the exercise in slide 70 where the mode mean and standard deviation are difficult to calculate.
tan
Mean Mode
Skewness
S dard Deviation
−
=
( )3
tan
Mean Median
Skewness
S dard Deviation
−
=
( ) ( )
( ) ( )
3 2 2 1
3 2 2 1
Q Q Q Q
Skewness
Q Q Q Q
− − −
=
− + −
1 2 3, ,Q Q Q 2Q
64. Moment Measure:
Where is the third moment and is the Standard deviation. For symmetric distribution, for each positive value of (xi-
mean) there is a negative value. When the deviations are cubed, positive values retain their positive sign and negative
sign and so m3 will be zero.
But for positive skews, large positive values of (xi-mean) are magnified considerably when cubed making m3 positive
and vice-versa for negative skews.
Thus positive, negative and zero values of correspond to positively skew, negatively skew or symmetrical curves.
( ) 3
1 3
m
Skewness γ
σ
=
3m σ
1γ
65. Kurtosis [22]: Two distributions may have the same measures for central tendency, dispersion and
skewness but the concentration of the values around the mode can be different. This concentration is called Kurtosis.
It defines the shape of a datasets distribution or otherwise speaks about the peak or flatness of a distribution as against the
normal distribution.
According to the peak of the distributions (the kurtosis) can be Leptokurtic curve with a maximum peak, Mesokurtic or a
normal curve with a normal peak and a Platykurtic curve with a flat peak as shown in figure below.
Kurtosis [31]
66. Measure of Kurtosis:
Where is the third moment and is the Standard deviation. For symmetric distribution, for each positive value of (xi-
mean) there is a negative value.
Thus positive, negative and zero values of correspond to leptokurtic, platykurtic and mesokurtic curves.
( ) 4
2 24
3 3
m
Kurtosis γ β
σ
= − = −
4m σ
2γ
67. Thus
Representative of a dataset is given by Central measures of Tendency
Variation / dispersion of Dataset from central tendency is given by Standard
Deviation
Direction of the distribution or the presence of data to the left or right of the mean
is given by Skewness
Concentration of data around the mode is given by Kurtosis
69. Exercise
Find the suitable measure of Skewness for the following distribution
SALES 0-20 20-50 50-100 100-250 250-500 500-1000
Firms number 20 50 69 30 25 19
70. Clues for Solution.
There are three formulae for calculation of Skewness:
Pearson’s first measure which uses Mode, Mean and Standard Deviation
Pearson’s second measure which uses Mean and Standard Deviation.
And the Bowley’s Measurement which uses The three quartiles.
The Mean, Median, Mode and Standard deviation involved in the first two formulae are difficult calculating for the above
grouped dataset and so the Skewness for this dataset is calculated using Bowley’s Measurement.
Here data items are in the groups (0-20, 20-50, 50-100,……, 500-1000), so for finding the Quartiles use the
cumulative frequency values of the data items.
Here Q1= value corresponding to n/4th Cumulative frequency
Q2= value corresponding to n/2th Cumulative frequency
Q3= value corresponding to 3n/4th Cumulative frequency
Where n is the total number of items (here the total number of firms =213) and the data items are sales values
from 0-20, 20-50,…
Q1= value corresponding to n/4th Cumulative frequency=213/4th value= 53rd sales value (from the table, 39 approximately )
Q2= value corresponding to n/2th Cumulative frequency=213/2th value= 106thsales value (from the table,76 approximately )
Q3=value corresponding to 3n/4thCumulative frequency=3x213/4thvalue=159thsales value (from the table, 203
approximately)
Use these Q1, Q2, Q3 values in Bowley’s Measurement and find the Skewness measurement.
SALES 0-20 20-50 50-100 100-250 250-500 500-1000
Firms number 20 50 69 30 25 19
Cumulative 20 70 139 169 194 213
frequency