This document discusses a study that examines whether randomness varies as a function of sample size. It describes 4 conditions with different sample sizes (n=3, 7, 30, 100) that were each subjected to 30 trials generating random numbers. Correlations between the random numbers were calculated for each trial. As sample size increased, the standard deviations of the correlations decreased, indicating the correlations became more tightly clustered around the predicted mean of 0, representing no linear relationship. The results supported the hypothesis that randomness is affected by sample size, decreasing as sample size increases, showing that as sample size grows, the outcomes more closely resemble what is predicted by the normal distribution according to the Central Limit Theorem. Throughout the study, only 7 type I errors occurred
The one page document was located at http://www.chirozone.net/TheDivineMatrix.pdf. I simply show the equation to illustrate the power of the caring individuals, "intenders" or "givers". Simplified states take the square root of the population intended to be affected by the intenders, then divide it by 10. For instance, if we want to affect 100 people, taking the square root produces 10 which divided by 10 results in one. Yes, the power of one rule. We can make a change in the world by simply believing that we can. See "The Power of Awareness" which can be summarized as follows:
The shadow effect enlightenment based upon âThe Power of Awarenessâ *
â it is only when we have the courage to face things exactly as
they are without any self-deception or illusion that the light
will develop out events by which the path to success may be
recognizedâ
I Ching
The one page document was located at http://www.chirozone.net/TheDivineMatrix.pdf. I simply show the equation to illustrate the power of the caring individuals, "intenders" or "givers". Simplified states take the square root of the population intended to be affected by the intenders, then divide it by 10. For instance, if we want to affect 100 people, taking the square root produces 10 which divided by 10 results in one. Yes, the power of one rule. We can make a change in the world by simply believing that we can. See "The Power of Awareness" which can be summarized as follows:
The shadow effect enlightenment based upon âThe Power of Awarenessâ *
â it is only when we have the courage to face things exactly as
they are without any self-deception or illusion that the light
will develop out events by which the path to success may be
recognizedâ
I Ching
Chapter 9 and 10 from book Understanding Research in SLL by James Dean Brown.
Chapter 9: Statistical Logic
Chapter 10: Correlation
Identifying a problem; Operationalizing Variables; Research Hypotheses
Choosing the correct statistic
Statistical hypotheses
Alpha decision level
Observed statistics
Assumptions:
1: Independence
2: Normal Distribution
3: Interval Scales
4: Linear Relationship
Degrees of Freedom
Critical Values
1. The nth Degree 17670556
1
Correlation to the nth Degree:
Does Sample Size Matter?
17670556
Psych. 4000
Dr. McGahan
October 20, 2015
-1
-0.75
-0.5
-0.25
0
0.25
0.5
0.75
1
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
Correlations for All Samples
n=3 n=7 n=30 n=100
2. The nth Degree 17670556
2
Correlation to the nth Degree:
Does Sample Size Matter?
Michael Guice
17670556
Psych. 4000
Dr. McGahan
October 20, 2015
-1
-0.75
-0.5
-0.25
0
0.25
0.5
0.75
1
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
Correlations for All Samples
n=3 n=7 n=30 n=100
3. The nth Degree 17670556
3
Abstract
The age old debate between freewill and determinism still stands today. Everything boils down
to essentially one question, who/what is in control? Some individuals see it as they are
completely in control of their lives, while others see it as be controlled by something else. For
freewill to exist there cannot be a predetermined causal chain. Using Karl Pearsonâs Product
moment correlation coefficient we try to determine whether randomness exist. Following these
lines, if randomness does exist, would it vary as a function of sample size? This idea permeates
as the theme of this study. Looking back at the Central Limit Theorem we see that normal
distributions tend to arise when an object is subjected to a larger number of independent
disturbances. This being said, a study has been designed to measure the effect of sample size on
randomness. Four conditions exist within this study. Each condition has a sample size associated
with it, and each condition was subjected to 30 trials a piece. When comparing the conditions we
used aggregate means to allow for greater understanding. The means showed us that as sample
size increased the standard deviations decreased. This means that the group became a tighter fit
around the predicted mean of 50. Randomness showed us its existence with a sample size of
100. It was predicted that as sample size increased we would fall more within the bounds of the
normal distribution. Although this may be true, condition 4 showed signs of chance.
Throughout the study only 7 type I errors were committed. This means seven null hypotheses
were incorrectly rejected.
4. The nth Degree 17670556
4
Correlation to the nth degree:
Does Sample Size Matter?
In one of Anthony Burgessâs more famous works, A Clockwork Orange, the protagonist
is put through a classical conditioning scenario. After the scenario is over Alex, the protagonist,
complains of how he had lost his freewill. What Alex was referring to was his inability to have
control over his actions. Merriam-Webster defines Free Will as the ability to make choices that
are not controlled by fate or God. Some believe that free will is a function of many âinputs,
including genetic and environmental factorsâ [Bradley, 2012]. This type of free will is known as
incompatibilist free will. Bradley referred to a coin flipping machine when mentioning this issue.
He said that if the machine had incompatibilist free will, âthe exact moment and the manner of
(a) release is inherently unpredictableâ. This unpredictability breaks a metaphorical causal chain
of events. The causal chain that locked Alex and his sickness together was the same chain that
removed his ability to steer his way through life. Without being able to steer Alex was more, or
less, dragged through life by something different than free will. But what was doing the
dragging? Alex believed it to be determinism.
Determinism
The most basic meaning of determinism is that only one course of events is possible.
James Bradleyâs work, âRandomness and Godâs Natureâ, elaborates on the various type of
determinism. In his work Bradley defines determinism as âthe philosophical position that
ontological randomness does not exist in the physical worldâ. Ontological randomness assumes
that randomness is a property of the very nature of things; as a side note, Bradley also mentioned
the idea of epistemic randomness (apparent randomness-a function of human perception of
things and not their nature). In comparison, ontological is âtrue randomnessâ and epistemic is
only random to the perceiver. With this being said, determinism would hold that free will is
imaginary from an epistemic viewpoint. Looking at Darwinâs Theory of Evolution, this would
suggest that evolution is a causal chain of predetermined events. If anything or anyone could
have complete control over any specific moment in life, the whole layout could change.
The idea of free will is the basic theme of Darwinâs Theory of Evolution. In this theory
Darwin stresses the idea that âthe most suited part will exist to reproduceâ. This idea delineates
from the idea of determinism in that there is the factor of âmost suitedâ. Going back to Alex, we
see that anytime he thought of doing something âbadâ a sickness feel over his entire body and
forced him to stop. This sickness created a disadvantage for Alex. Without control Alex
believed he could not be held responsible for any action he did, or did not, commit. At this point
he began to believe his fate was sealed. This deterministic mindset forced Alex into a depression
and later an attempted suicide. The very people who âhelpedâ the protagonist later tried to
dispose of him. **Spoiler Alert** Just so happens luck was on his side and he lived through it.
The question is, was the luck predetermined or was it a matter of âblind, purposeless chanceâ
[Bradley, 2012]?
Randomness
Chance is usually mentioned in the study of randomness and proportionality. A popular
conception of randomness is ânot having a governing design, method, or purpose; unsystematicâ
[Bradley, 2012]. A fair die has six sides and when thrown has the probability of landing on one
of those six sides. Much like the results in Neuringerâs study, âhumans failed to produce random
5. The nth Degree 17670556
5
like-behaviorâ, die tend to fail to produce random like-behavior. For instance, when an
individual plays a game of chance and rolls a die they expect one-out-of-six possibilities. What
happens when the die balances on a corner? What if the die rolls of the table and disappears
forever? These questions give way to randomness. The question proposed now is, does
randomness function as sample size fluctuates? Will the die produce any response other than
one of the six possibilities mentioned before? Seeing as those chance is a function of
randomness this study was aimed at determine whether randomness existed, and if so does it
vary as a function of sample size.
Correlation and Prediction
The Pearson product-moment correlation allows researchers the ability to detect linear
relationships. In order to use the method three assumptions must be meet;
1.) âThe Sample is independently and randomly selected from the
population of interest.â
2.) âThe population distributions of X and Y are such that their joint
distribution (that is, their scatterplot) represents a bivariate normal
distribution. This is called the assumption of bivariate normality and
requires that the distribution of Y scores be normal in the population at
every value of X.â
3.) âThe variances of the Y scores are equal at every value of X in the
population.â
What does a correlation coefficient do? A correlation coefficient measures the strength
of the association between two variables if there is one. âThe Pearson product-moment
correlation coefficient measures the strength of the linear association between
variables.â[StatTrek]. The coefficient bounds are between negative one and positive one. Using
this scale allows the researcher to depict how influence on one variable may affect the other. A
correlation of negative one should represent a perfect negative-relationship between the two
variables. Variable one should move incrementally away from variable two when two is
manipulated, and vice versa. When two variables move sequentially in the same direction when
the other is adjusted we would consider this a positive relationship. Correlation coefficients
equal to zero do not represent âzero relationship between two variables; rather, it means zero
linear relationshipâ [StatTrek]. What is a linear relationship?
A linear relationship is a relationship of proportionality. Generally this relationship is
plotted on a graph and the points create a straight line. Any change in one variable, the
independent variable, will emit a corresponding change in another variable, the dependent
variable. Practice makes perfect. As an individual spends more time practicing their ability
should increase within that activity. Consequently, by reducing the time spent practicing oneâs
ability should diminish as well.
Galton:
The concept of correlation was first discovered by Francis Galton in 1888. This came
after his intensified research on heredity. Galton wanted to âreconcile an empirical test with a
mathematical theoremâ [Stigler, 1989]. Galton used a Quincunx, an ingenious analogue
computer, to help him with his correlation formulations. Although Galton used median
deviations from the median, today we âexpress measurements in terms of the number of standard
deviations units from the meanâ [Stigler, 1989]. Although Galton is credited with the discovery
6. The nth Degree 17670556
6
of correlation, Karl Pearson is credited with the discovery of the Pearson r product-moment
correlation.
Pearson:
Before Pearsonâs discovery, Venn diagrams were used to determine linear relationships.
Karl Pearson was quoted saying that ârecess deserves a commerative tablet as the birth place of
the truce conception of correlationâ [Sigler, 1989]. Itâs believed that Pearson meant no
disrespect towards Galton. More simply stated, the idea of correlation did not âclickâ for Galton
until he took time to reflect.
Pearson was raised as a Quacker until the age of 24. While Pearson studied under Edward
Routh at Cambridge he began to lose his religious faith. Itâs been said he began to adhere to
agnosticism or âfreethoughtâ [Brittancia]. Freethought holds that positions regarding truth
should be formed on the basis of logic, reason, and empiricism. This idea is directly negatively
correlated with the idea of authority, tradition, revelation or other dogma. Once the correlational
method became an integral part of science it opened the door for a new zeitgeist.
Intelligence Testing and General Model of Reliability
Intelligence Testing:
Intelligence testing was first introduced by Lewis Terman in 1916. Terman was a
psychologist at Stanford University. At the time the test was administered to two year old and
older participants. These test were individually administered and consisted of âan age-graded
series of problems whose solutions involved arithmetical, memory, and vocabulary skills.â
[Britannica]
The most popular intelligence tests are the Stanford-Binet Intelligence Scale and the
Wechsler Scales. The Stanford-Binet is the American adaptation of the original French Binet-
Simon intelligence test. An IQ, or an intelligence quotient, is a concept first suggested by
William Stern. It was originally computed as the âratio of a personâs mental age to his
chronological (physical) age, multiplied by 100â. [Britannica] Today the mental age has fallen
by the wayside. Test results still yield an IQ, but the concept is now configured on the basis of
statistical percentage of people who are expected to have a certain IQ.
Intelligence test scores follow an approximately ânormalâ distribution. This normal
distribution suggest most people score near the middle of the distribution curve. As one diverts
attention away from the mean scores tend to drop off fairly rapidly in frequency. Even though
intelligence testing seems simple there is still room for error.
Errors:
There are two types of errors that plague the scientific community, type 1 and type 2.
Type 1 error is seen as incorrectly rejecting the null hypothesis. The scientific community is
protected against over hastily rejection of the null hypothesis by statistical testing. Although
statistics are never 100% accurate this allows âsome policing powers to members who would
rather live by the law of small numbersâ [Kahneman and Tversky, 1971]. Kahneman and
Tversky went on to state âthere are no comparable safeguards against the risk of failing to
confirm a valid research hypothesis (Type II Error).
Replication
7. The nth Degree 17670556
7
For this study samples of 30 will be collected. This idea goes back to the article
âProbable error of a correlation coefficientâ published in 1908. At the time William Gosset was
publishing under the pseudonym Student. William, head brewer of Guinness at the time,
proposed the idea that âwith samples of 30âŚthe mean value (of the correlation coefficient)
approaches the real value (of the population) comparatively rapidlyâ. Why should we replicate
in the first place?
Kahneman and Tversky (1971) stated that âthe decision to replicate a once obtained finding
often express (es) a great fondness for that finding and a desire to see it accepted by a skeptical
communityâ. In this same research, Kahneman and Tversky showed that 88% of this skeptical
community believed the results of a single significant results are likely do to chance. This idea
falls directly in line with the Central Limit Theorem (CLT). Stigler, 1989, referenced the CLT
and said âthat the normal distribution arises when an object is subjected to a larger number of
independent disturbances, no few of them dominantâ.
Design to Test
One study with four conditions has been developed. The study requires the use of âtrue
random numbersâ. These numbers will be retrieved from a random generator, Random.org.
Unlike pseudo random numbers, created from algorithm, Random.org generates numbers from
atmospheric noise. Conditions were separated by sample size;
Condition 1 Sample Size of 3
Condition 2 Sample Size of 7
Condition 3 Sample Size of 30
Condition 4 Sample Size of 100
As the conditions change so do the sample size. This is due to Kahneman and Tverskyâs
idea that âa replication sample should often be larger than the originalâ.
Variables X and Variable Y were pooled from the environment. Due to the nature of
randomness variable X should not predict the Variable Y. Variable Y should not predict
Variable X. This being the case Variable X and Variable Y should not be highly correlated.
When correlated the Pearson-product moment correlation should be non-significant at an alpha
level of .05. This alpha level was selected due to its ability to produce a confidence level of 95%.
This is also the most prominent alpha within the scientific community. Also, the aggregate mean
of the Pearson-product moment correlations between the samples should not exceed the critical
values. Critical values for the samples are listed below.
Sample Size: Critical Value: Degrees of Freedom
(two-tailed test)
N=3 0.997 1 (3-2=1)
N=7 0.754 5 (7-2=5)
N=30 0.361 28 (30-2=28)
N=100 0.197 98 (100-2=98)
Degrees of freedom are essential to this research and thus it was added to the table of
critical values.
Degrees of freedom and accuracy of estimates tend to be positively correlated. As the
sample size increases the degrees of freedom increase. With this the accuracy of the estimate,
the aggregate Pearson-moment correlation means, should be more closely related to the
population mean. As the sample sizes increase the aggregate means should group more tightly
around zero, with zero being understood as representing no linear relationship. This in turn
should be representative of smaller standard deviations. Each conditionâs standard deviation
8. The nth Degree 17670556
8
ârepresents an average deviation from the meanâ [Jaccard, J., 2010]. In theory as sample size
increases one should be better able to predict the effect size of the variable in question. For this
study the question does randomness function as a variable of sample size, has been posed. The
hypothesis for this study is that randomness does vary as a function of sample size.
To test this hypothesis random numbers will be generated by Random.org. Two
variables will be drawn for each subject within the conditions and formatted into two columns.
Column1 will thus be denoted as Variable 1 and Column2 will be denoted as Variable 2. Each
condition should use a multiplier of 2 for each participant. This means there will be a total of six
variables for a sample size of three, fourteen for a sample size of 7, sixty for a sample size of 30,
and two hundred for a sample size of 100.
Each condition should have thirty trials. The first initial trial and twenty-nine
replications. At the end of each trial a test of correlation will be ran to determine the relationship
between the two variables. While testing for the correlation, standard deviations and means for
each variable should be collected for later comparison. Once all 30 trials have been ran for each
condition the Pearson correlation results should be averaged to determine the grand mean of the
samples. âThe mean of a sampling distribution of the mean will always be equal to the
population mean (of the raw scores)â [Jaccard, J., 2010]. This process should also be completed
for the standard deviations, as well as the means, of Variable1 and 2 for all conditions.
Sample Size
As mentioned before, sample size is a key variable for this study. In fact, it is the only
variable and thus is labeled as the independent variable (IV).
N=3
Ever heard either of the following sayings, âThe third times the charmâ or âAll good/bad
things come in threesâ? Three seems to be a special number when it comes to people in general,
but especially in religion. For instance, Christians tend to symbolize the Holy Trinity in terms of
three (Father, Son, and Holy Spirit). Even Godâs attributes tend to be focused in threes
(omniscience, omnipresence, and omnipotence). The Christian faith is not the only faith where
this trend of threeâs appears. Taoist believe in the Great Triad which includes heaven, human
and Earth. Throughout the research 3âs and religion go hand and hand. Even in everyday life
the number three pervades our minds (past, present, and future). Due to its evasive presence in
the natural, and supernatural, world condition 1 will include a sample size of three. It only
seems appropriate to study natural evasiveness when researching randomness. Also, this
particular sample size will produce one degree of freedom. If the accuracy of a prediction varies
as a function of degrees of freedom, then condition1 should have aggregate means much
different than the other three conditions. This idea is supported by the Law of Large Numbers,
the more samples you use the closer your estimates will be to the true population values. Only
when a self-correcting tendency is in place should a small sample be highly representative and
similar to that of a larger sample [Kahneman and Tversky, 1971]. Due to the nature of our data,
self-correcting tendencies are nor, and should not be in place. Thus condition 1 should show the
largest variance in comparison to the population so long as the other conditions represent larger
samples sizes.
N=7
The number 3 is not the only significant number that appears throughout religion and our
daily lives. It took seven days for God to create the Earth. âSeven is the second perfect numberâ
9. The nth Degree 17670556
9
only losing to three. [Scripture] In the book of Revelation seven makes thirty-five (5 x 7 = 35)
appearances. Similarly, âseven is a significant number in the natural world: mammals and birds
have a gestation of multiples of 7â [Scripture]. These issues alone make seven seem like an
attractive number to include within a study of randomness. Then thereâs Millerâs âMagical
Number of Seven, plus or Minus Twoâ. Millerâs research concluded that a personâs memory
span is limited to seven chunks, plus or minus two. For these reasons condition2 has been
identified as the sample size of seven.
N=30
A sample size of thirty was particularly interesting for this study. In âGod Countsâ W.E.
Filmer translated Bible verses into its native tongue, Hebrew. After doing so he associated a
numerical value to an idea. This is known as Bible numeric. âAs each idea is introduced the
associated number echoes throughout all manner of divisions and classifications in a way which
cannot be put down to mere chanceâ [Filmer]. This research became relevant with further
investigation of the value of thirty. Filmer shows us that 30 seems to representative of the idea
of âBlood of Christ; Dedicationâ. What was really cool was the degrees of freedom for a sample
size of thirty. A sample size of thirty has twenty-eight degrees of freedom. The idea associated
with twenty-eight degrees of freedom is âEternal Lifeâ, according to Filmer. A sample size of
thirty was a must-have for condition 3.
N=100
For the fourth condition the study required a sample size that was significantly larger than
any of the other three. Looking back to the Law of Large numbers, Kahneman and Tversky
stated that â(the law) guarantees that very large samples will indeed by highly representative of
the population from which they are drawnâ [1971]. Whatâs larger than 100%? Although we can
never really achieve 100% in almost anything in real life, the idea is attractive. In fact, scientist
have acknowledged the fact that we are unable to obtain a 100% accurate finding due to minimal
levels of error. Another strange factoid, 100 is 3.33 the size of condition 3 (all those threes!!).
For these reasons condition 4 is identified as sample size of 100.
Method
Rather than running the correlation by hand I chose to test the data with statistical
software. So long as the data was entered correctly and the correct boxes where checked,
statistical software has a better chance of acquiring the correct outcome. Secondly, the software
saves valuable time.
Generating random numbers.
As mentioned before, random.org generates random numbers through the use of âtrue
randomnessâ [random.org]. Using this websites features allows the ability to test whether âtrue
randomnessâ exist. By typing the web address into the URL, ârandomness.orgâ, and initiating
the search, a webpage should appear. Welcome to Random.org! Now for the numbers that you
will be using for data.
Near the top of the webpage there should be a strand of blue, hyperlink, words. The one
we are most concerned with today is the hyperlink entitled âNumbersâ. Scroll over this
10. The nth Degree 17670556
10
hyperlink and a drop box should appear. The drop box has a list of choices in the following
order; integers, sequences, integer sets, Gaussian numbers, decimal fractions, and raw bytes.
The choice most relevant to this study is âintegersâ. Reason being is you want plain data. Now,
click âintegersâ.
When the new page is finished loading you will notice you are now looking at the
âRandom Integer Generatorâ. This generator will âallow you to generate random integersâ that
are produced by atmospheric noise. [random.org] before you begin you should become familiar
with the layout of the webpage. There are two very important sections to this site; Part 1: The
Integers and Part 2: Go! Part 1 is where you inform the generator how many integers you want
to produce, the range in which you wish the numbers to be derived, and how many columns you
want the generator to format your data.
For the first condition, sample size of three, the generator should produce 6 random
integers with a value between 0 and 100. This scale has been used because students tend to
understand this scale seeing as though most grading rubrics have a range of 101, 0-100. Also,
the data should be formatted into two columns. By doing so, we produce three samples with two
variables each. Column one should be recognized as variable one and column two as variable
two. Once the form has been filled out correctly the generating can begin! Advert your attention
to Part 2: Go! At this point three options are available; get numbers, reset form, switch to
advanced mode. For this experiment the first option, Get Numbers, will suffice. Option 2 will
clear all your hard work thus far and opinion three is
beyond the scope of this particular topic. Now generate.
As your generated numbers appear you will see two
columns with three rows each. For instance, if you
generated a set of numbers that look like table 2.1 you did
it correctly. What this table tells its audience is that
subject 1 had a variable 1 equal to 27. Likewise, subject 3
has a response of 14 for its variable 2. The pattern
continues as you follow the flow of the table. In its entirety, table 2.1 represents trial 1 for a
sample size of 3. For each condition in this study 30 trials will be ran for each sample size.
Returning to the website, near the bottom left of the webpage there are two options.
Option 1(Again!) and option 2 (Go Back). The quota for the condition has not been met. 29
trials still remain, thus making option 2 an unlikely candidate for selection. With that being said,
Option 1 should be selected. This action will provoke the generator to produce six more integers
in the same format. Now for the other three conditions.
The other three conditions require a sample size of 7, 30 and 100 be used respectively.
The same process that is listed above for a sample size of 3 will be used with a few exceptions.
The range and the number of variables (columns) are controlled within the respective confounds
of this experiment. These controls allow the researcher to pin point the relationship between
variability and sample size. For a sample size of 7, fourteen integers will be generated. These
fourteen integers will be divided into two separate columns providing two variable responses for
seven subjects. This method is continued for both sample sizes of 30 and 100. For the sample
Table 2.1
Variable 1 Variable 2
Subject 1 27 37
Subject 2 92 5
Subject 3 16 14
11. The nth Degree 17670556
11
size of 30, sixty responses will be required. You guessed it! Two hundred responses will be
required for the sample size of 100. Remember!! Each condition requires thirty trials. Now
what do with this generated data?
Each data set should be placed into a safe place for keeping and further study. For this
study we have chosen to insert our data into SPSS, Statistical Package for the Social Sciences,
software. This program has been chosen due to the nature of the data and the high utility of the
software. Itâs worth mentioning that version 22 of SPSS was used for this particular study. Upon
opening the SPSS software the user is met with two windows. The first of which inquires about
the userâs need, and the second should be the dataset viewer page. On the first window, near the
top left you should double-click âNew Datasetâ underneath the New Files tab. By doing so, the
software program will present a document usually entitled âOutput 1â. This is where all of your
formulations and equations will send their answers. For now, letâs put our attention on the other
window now available, âData Editorâ.
Once inside this window you should notice various amounts of columns and rows. The
rows should be numbered in an ascending manner, and the columns should be labeled âvarâ. For
instance, column will be list as âVar1â. Also, near the bottom left hand side of the window there
should be two tabs. The tabs should be labeled âData Viewâ and âVariable Viewâ. Variable
view allows the user to determine the scale of measurement for each variable. If there are any
concerns on this matter please see S.S. Stevensâs article entitled âOn the Theory of Scales of
Measurementâ. In this article Stevens refers to the four scales of measurement, NOIR, and how
each is used. For the moment any further elaboration is beyond the scope of this study. The
âData Viewâ is where the user will spend most of their time. As mentioned before, there are two
variables for each trial and this theme permeates throughout the entire experiment.
Inserting the Data:
After generating the numbers in the random generator one must transpose these numbers
into the SPSS software. Column 1 created by the random generator will be inserted into the first
available odd-numbered column inside SPSS. For instance, if VAR1 and VAR5 were both
empty, VAR1 would be the home for the first column of the first round of data. VAR2 would be
the home for the second column of generated data. Each column of generated data should be
entered consecutively. The goal is to create sixty VARs within each condition. This decision is
based on the idea that each condition requires 30 trials, and each trial requires 2 sets of variables.
As the sample size increases the amount of variables, VAR, do not increase; only the amount of
participants under each column heading will do so. These steps have been take to ensure
accuracy within the calculations, as well as to avoid any confusing results.
Churning the Data:
All four conditions, sample size of 3,7,30, and 100, will be analyzed through SPSS.
Although each condition must be performed in the same manner, each must be performed
separately. This protects against the misrepresentation of data and later confusions as to the
results.
12. The nth Degree 17670556
12
Once all the data has been inserted and saved into its respectable files the data can then
be âchurnedâ, or analyzed. Near the top-left corner of the SPSS window there is an option
called âFileâ. This particular item is unimportant for the time being, but its location is pivotal in
a manner of speaking. To the right of âFileâ should be âEditâ. Continue to look further right
and you will come across an item referred to as âAnalyzeâ. Remember the location for this
selection. It is a key path way to analyzing your data. Now select âAnalyzeâ. A scroll down bar
should pop up and the first item for selection should be âReportsâ. Continue to look further
down until you see the selection âCorrelateâ. Upon placing the cursor over the âCorrelateâ
function another drop box appears. Select the option âBivariateâ. This option has been selected
because our data involves two variables; variable 1 and variable 2. After selecting âBivariateâ a
new window entitled âBivariate Correlationsâ should pop-up. This is where you will select the
variables you want to âchurnâ.
The left side of the new window contains all the variable you have entered. Again, for
each condition there should be 60 variables listed. By highlighting the variable in the left
column and either double-clicking, or manually inserting with the arrow, you can turn your
available variables into variables of interest. Once the two variables, VAR1 and 2 (or VAR9 and
10), have been successfully been transformed in the right side you can begin to inform the
program what it is you need. The âOptionsâŚâ tab allows the user to add other useful data inside
the output. This data may be beneficial if selected and depicted in the proper manner. For this
experiment âmeans and standard deviationsâ underneath the statistics heading should be selected.
Once this has been selected you can select âContinueâ near the bottom of the page. This will
send you back to the âBivariate Correlationsâ page. While in this page be sure to select the
âPearsonâ box under Correlation Coefficients. The lets the program know you are looking for
the Pearson r moment-correlation, as mentioned before. The test of significance should also
have the option âtwo-tailedâ selected. Lastly, be sure to check the box âFlag significant
correlationsâ. This will make you work much easier while looking at the outputs. Once all the
necessary steps have been completed select âOKâ at the bottom of the page. This process will
need to be completed 30 times for each condition. After the first correlation is ran there is no
need to go through the options selection again. SPSS saves your selection choices so long as you
do not exit the program. The key thing to remember when running these analyses is the variable
must be replaced with each replication. For instance, VAR 1 and 2 must be removed from the
right side so VAR 3 and 4 can be analyzed.
SPSS: âOutputâ
After selecting âOKâ in the bivariate correlations page the program should direct
attention to the output window. This is where the churned data can be analyzed and viewed.
Some numbers of interest on this page include; the mean of both VAR1 and VAR2, the standard
deviation for both variables, and the Pearson correlation for VAR1 as compared to VAR2. The
Pearson correlation for VAR2 as compared to VAR1 should be identical due to the mirror effect
of the matrix. These numbers should be entered and saved into a Microsoft Excel workbook
page. Data labels for each condition should be five columns wide. Underneath these labels there
should be five column labels listed horizontally as follows; âr, Mean V1, Std. Dev V1, Mean V2,
and Std. Dev V2â. Along the vertical axis of the workbook should be the word âTrial (with its
13. The nth Degree 17670556
13
respective number of 1-30)â. The data that will be inserted into this workbook will be derived
from the previous application in SPSS.
Excel: Tables and Charts
Visual representations have the potential to help the viewer follow the flow of the data
and allow them to observe its physical nature. Once the new data is plugged into Excel it can
now be used to create visual representations of the data. Particular visual of interest for this
study include; Pearson r scatter plots, Pearson r line charts, sample means line chart from each
condition, as well as the sample means scatter plots.
A Pearson r scatter plot can be created by highlighting the correlation data for all four
conditions and selecting the option âInsertâ. This option can be located at the top of the Excel
window to the right of âHomeâ. Once the tab has adjusted items such as âRecommended
Chartsâ, âPivot Chartâ, and âTablesâ should be visible. There should also be a diagram eluding
to that of a scatterplot. The diagram appears to have an X and Y axis with small dots spatially
organized. If this option does not appear click any image of a chart and redefine your selection
on the right side of the pop-up window; if it does, click it. This new window allows you to
manipulate the visual aspects of the chart itself. Pick the best way to represent your data thatâs
not misleading or uninformative.
Pearson r line charts have the potential to be confusing. This may be true because all four of
the conditions
correlations are laid on-
top-of one another. If
there is lots of variation
between the conditions
this chart has the potential
to be confusing.
Nonetheless, the
confusion may allow the
user the potential to
differentiate between the
conditions. This type of
chart can be created by
highlighting all the variables of interest and inserting them into a line chart the same way the
scatter plot was completed, but by selecting line chart instead of scatter plot.
Results
For each condition one chart has been selected to represent both variables. Var1 for each
condition will be used for a visual reference point.
-1
-0.75
-0.5
-0.25
0
0.25
0.5
0.75
1
1 2 3 4 5 6 7 8 9 101112131415161718192021222324252627282930
Correlations for All Samples
n=3 n=7 n=30 n=100
14. The nth Degree 17670556
14
The results of this study show that for condition 1,
n=3, the aggregate means for Var1 and Var2 are
47.4332 and 49.9444, respectively. Similarly the
aggregate mean for the standard deviation of Var1
is 20.16363 and Var2 is 19.26425. The grand
mean for the ranges for Var1 and Var2 are 83.33
and 71.00, respectively. The chart entitled
âVar00001â resembles the general sample of
condition one. Notice how the x-axis ranges from
0-100. Also notice the shape of a ânormal bell
curveâ and how the histogram does not exactly
âfitâ.
For condition 2, n=7, the aggregate
means of Var1 and Var2 are 50.0097 and
49.7048, respectively. Aggregate
mean of the standard deviation for Var1
was 10.39435, while Var2 was 9.06379.
The two grand means for the ranges were
38.43 and 35.86, respectively. The chart
entitled âVAR00003â represents Var1
from condition 2. Be sure to notice the
x-axis is no longer bound to 0-100, but
rather 20-70. This reveals the tightening
of the bounds. Also, be aware of the
tighter grouping as compared to the
graph illustrated in condition 1. As the
standard deviation begins to decrease the
means begin to fall under the normal distribution.
Condition 3, n=30, aggregate means of Var1 and
Var2 were 49.9722 and 49.8778. The aggregate
mean of the standard deviations were 5.11790 and
5.32132. The grand means for range equaled
19.77 and 20.17. The graph entitled âVAR0005â
represents Var1 from condition 3. Notice once
again how the x-axis bounds have decreased. The
bounds are now restricted between 40-65. As the
sample size increases more of the means tend to
fall under the normal distribution.
Lastly, condition 4, n=100, the aggregate means
were 48.9833 and 50.0453. The aggregate means
of the standard deviations were 2.66304 and
2.58452, respectively. The grand means of range
15. The nth Degree 17670556
15
equaled 9.71 and 10.09. The graph
entitled âVAR00007â represents the
Var1 from conditon 4. Yet again as
sample size decreases the bounds of the
x-axis have decreased. However, unlike
the other graphs to this point, this graphs
means tend to be more spread out from
the curve.The pattern till now was a
tighter-more compressed grouping. It
would appear as though this sample size
was effected by chance.
Correlation Results
Correlation Results
Critical
Values
Mean Standard
Deviation
Range
Condition 1 0.997 -.1000 .60864 1.93
Condition 2 0.754 .0248 .49867 1.59
Condition 3 0.361 -.0344 .18345 .72
Condition 4 0.197 .0009 .12054 .43
The table entitled âCorrelation Resultsâ shows the correlation results for this study. The
columns entitled âMeanâ, âStandard Deviationâ and âRangeâ should all be seen as aggregate
means. The means were taken from the 30 trials within each condition and collapsed into one
average. By doing so, we have saved valuable time and space for both the reader and researcher
alike. This table shows that on average none of the conditions grand mean were greater than the
critical values associated with each sample size. Thus no significant results were found to
disprove that randomness existed. Similarly, as predicted, as the sample size increased
throughout the conditions the aggregate mean of the standard deviation shrank incrementally.
Likewise, the range followed suite. On the contrary itâs worth mentioning that although the
aggregate mean of the correlations did not exceed the critical values, some individual trials did.
For the entire study only seven type I errors were made. These errors are known due to
the nature of the data. The data is random data and should not portray a significant correlation as
mentioned before. In the table below I have listed the amount of type I errors per condition;
Condition: Number of Type I Errors:
Condition1 1
Condition 2 4
Condition 3 1
Condition 4 1
Total Type I Errors 7
16. The nth Degree 17670556
16
Scatterplots for the Pearson r Product-Moment correlations for this study are shown and
described below.
This image has made an appearance multiple times throughout this study. Now itâs time to
dissect it. The blue line, n=3, shows the variation in the correlations. Notice how it seems to
take greater leaps âof faithâ as in crosses the flat horizontal line. The horizontal line can be seen
as 0, or having no linear relationship. In all the chaos it seems as though only two lines know
exactly where it is they are headed, the grey and yellow line. These lines represent condition 3
and 4 respectively. Although they both seem to be navigating quite well through the chaos,
condition 4 seems to hug the horizontal line a little tighter. This would fall right in line with the
idea that two random variables should not significantly correlate with one another. As sample
size increases, the degrees of freedom increase. When this happens it makes it easier for
statisticians, researchers, and students alike to predict linear relationships.
Conclusion:
The age old debate between freewill and determinism is not a simple cut-and-dry
scenario. If it were it probably would have been resolved long ago. The monkey wrench being
thrown into the mix is something called randomness. Randomness behaves in a spurious way. It
can be said that randomness is a function of chance. In this study we used Pearson product
moment correlation to determine the effect of sample size on randomness. Four conditions were
created with various amounts of samples. Sample size was the independent variable for this
study. Thus it should be noted that condition 1 had a sample size of 3, condition 2 had a sample
size of 7, condition 3 had a sample size of 30, and condition 4 had a sample size of 100. Another
key to this research was the amount of replications. Kahneman and Tverskyâs advice from the
law of small numbers was used to help predict our outcomes. We predicted that as sample size
increased we would be more able to predict the nature of randomness. This turned out to be
false. As sample size increased we were able to predict to a certain degree plus a minus a certain
degree, but we were never able to predict the next number to be generated. Ideas for future
research include, but are not limited to a sample size of 1000. With a sample size of 1000 it is
-1
-0.75
-0.5
-0.25
0
0.25
0.5
0.75
1
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
Correlations for All Samples
n=3 n=7 n=30 n=100
17. The nth Degree 17670556
17
believed that the aggregate means of the standard deviations will be relatively small in
comparison to our data.
18. The nth Degree 17670556
18
References:
http://www.britannica.com/biography/Karl-Pearson
http://www.britannica.com/science/intelligence-test
Bradley, J. (2012). Randomness and Godâs Nature. Perspectives on Science and Christian Faith,
Vol. 64, 2, 75-89.
Filmer, W.E. (1984) God Counts: Amazing Discoveries in Bible Numbers.
Jaccard, J. Becker, M. (2010). Statistics for the Behavioral Sciences. Cengage Learning.
Belmont, CA.
Neuringer, A. (2002). Operant Variability and the Power of Reinforcement. The Behavior
Analyst Today, Vol. 10, 2, 319-343.
Stigler, S. (1989). Francis Galtonâs Account of the Invention of Correlation. Statistical Science,
Vol 4, No. 2, 73-86.
S.S. Stevens. (1946). On the Theory of Scales of Measurement. Science, Vol. 103, 2684, 677-
680.
Tversky, A. Kahneman, D. (1971). Belief in the law of small numbers. Psychological Bulletin, 2,
105-10.
âScriptureâ
The Significance of Threes (1988). Agape Bible Study.