An introduction to mediation analysis using SPSS software (specifically, Andrew Hayes' PROCESS macro). This was a workshop I gave at the Crossroads 2015 conference at Dalhousie University, March 27, 2015.
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 12: Analysis of Variance
12.1: One-Way ANOVA
Introduces and explains the use of multiple linear regression, a multivariate correlational statistical technique. For more info, see the lecture page at http://goo.gl/CeBsv. See also the slides for the MLR II lecture http://www.slideshare.net/jtneill/multiple-linear-regression-ii
Please Subscribe to this Channel for more solutions and lectures
http://www.youtube.com/onlineteaching
Chapter 12: Analysis of Variance
12.1: One-Way ANOVA
Introduces and explains the use of multiple linear regression, a multivariate correlational statistical technique. For more info, see the lecture page at http://goo.gl/CeBsv. See also the slides for the MLR II lecture http://www.slideshare.net/jtneill/multiple-linear-regression-ii
This presentation explains the procedure involved in two-way repeated measures ANOVA(within-within design). An illustration has been discussed by using the functionality of SPSS.
My attractive effective presentation is the proof of my hard work as i made it for those who can not take interest in their studies so as they can see this they will take interest too as well as for those who really want to do come thing different from others , they can use my presentation if any kind of help you want just mail me at ammara.aftab63@gmail.com
This is the basic explanation on what are ANCOVA and MANCOVA in research study in which provides the definitions and the illustration on how can these both be use in SPSS tool analysis. If you's like to get practice file, do not hesitate to contact me.
This presentation explains the procedure involved in two-way repeated measures ANOVA(within-within design). An illustration has been discussed by using the functionality of SPSS.
My attractive effective presentation is the proof of my hard work as i made it for those who can not take interest in their studies so as they can see this they will take interest too as well as for those who really want to do come thing different from others , they can use my presentation if any kind of help you want just mail me at ammara.aftab63@gmail.com
This is the basic explanation on what are ANCOVA and MANCOVA in research study in which provides the definitions and the illustration on how can these both be use in SPSS tool analysis. If you's like to get practice file, do not hesitate to contact me.
Most data scientists are focused on predictive (aka supervised) projects, yet the real growth is usually in the estimation of action effects and optimizations of action policies. To this end, I will present causal inference and related packages.
There are three layers of analytics: descriptive (BI), predictive (supervised modeling), and prescriptive - the latter, the less-known one, focus on answering the most important business questions. For example, "what was the effect of giving a discount" ( or "what should I do to create the desired effect" - In this talk, we will first discuss what frameworks are used to answer these questions, namely causal inference, and reinforcement learning. Then we will deep dive into CI and discuss in causality crash 101 courses why is it important. Last but not least we will present existing causal-inference open-source packages and their limitations.
Between Black and White Population1. Comparing annual percent .docxjasoninnes20
Between Black and White Population
1. Comparing annual percent of Medicare enrollees having at least one ambulatory visit between B and W
2. Comparing average annual percent of diabetic Medicare enrollees age 65-75 having hemoglobin A1c between B and W
3. Comparing average annual percent of diabetic Medicare enrollees age 65-75 having eye examination between B and W
4. Comparing average annual percent of diabetic Medicare enrollees age 65-75 having
Students will develop an analysis report, in five main sections, including introduction, research method (research questions/objective, data set, research method, and analysis), results, conclusion and health policy recommendations. This is a 5-6 page individual project report.
Here are the main steps for this assignment.
Step 1: Students require to submit the topic using topic selection discussion forum by the end of week 1 and wait for instructor approval.
Step 2: Develop the research question and
Step 3: Run the analysis using EXCEL (RStudio for BONUS points) and report the findings using the assignment instruction.
The Report Structure:
Start with the
1.Cover page (1 page, including running head).
Please look at the example http://www.apastyle.org/manual/related/sample-experiment-paper-1.pdf (you can download the file from the class) and http://www.umuc.edu/library/libhow/apa_tutorial.cfm to learn more about the APA style.
In the title page include:
· Title, this is the approved topic by your instructor.
· Student name
· Class name
· Instructor name
· Date
2.Introduction
Introduce the problem or topic being investigated. Include relevant background information, for example;
· Indicates why this is an issue or topic worth researching;
· Highlight how others have researched this topic or issue (whether quantitatively or qualitatively), and
· Specify how others have operationalized this concept and measured these phenomena
Note: Introduction should not be more than one or two paragraphs.
Literature Review
There is no need for a literature review in this assignment
3.Research Question or Research Hypothesis
What is the Research Question or Research Hypothesis?
***Just in time information: Here are a few points for Research Question or Research Hypothesis
There are basically two kinds of research questions: testable and non-testable. Neither is better than the other, and both have a place in applied research.
Examples of non-testable questions are:
How do managers feel about the reorganization?
What do residents feel are the most important problems facing the community?
Respondents' answers to these questions could be summarized in descriptive tables and the results might be extremely valuable to administrators and planners. Business and social science researchers often ask non-testable research questions. The shortcoming with these types of questions is that they do not provide objective cut-off points for decision-makers.
In order to overcome this problem, researchers often seek to answer o ...
BUS 308 Week 5 Lecture 3 A Different View Effect Sizes .docxcurwenmichaela
BUS 308 Week 5 Lecture 3
A Different View: Effect Sizes
Expected Outcomes
After reading this lecture, the student should be familiar with:
1. What effect size measures exist for different statistical tests.
2. How to interpret an effect size measure.
3. How to calculate an effect size measure for different tests.
Overview
While confidence intervals can give us a sense of how much variation is in our decisions,
effect size measures help us understand the practical significance of our decision to reject the
null hypothesis. Not all statistically significant results are of the same importance in decision
making. A difference in means of 25 cents is more important with means around a dollar than
with means in the millions of dollars, yet with the right sample size both groups can have this
difference be statistically significant.
Effect size measures help us understand the practice importance of our decision to reject
the null hypothesis.
Excel has limited functions available for us to use on Effect Size measures. We generally
need to take the output from the other functions and generate our Effect Size values.
Effect Sizes
One issue many have with statistical significance is the influence of sample size on the
decision to reject the null hypothesis. If the average difference in preference for a soft drink was
found to be ½ of 1%; most of us would not expect this to be statistically significant. And,
indeed, with typical sample sizes (even up to 100), a statistical test is unlikely to find any
significant difference. However, if the sample size were much larger; for example, 100,000; we
would suddenly find this miniscule difference to be significant!
Statistical significance is not the same as practical significance. If for example, our
sample of 100,000 was 1% more in favor of an expensive product change, would it really be
worthwhile making the change? Regardless of how large the sample was, it does not seem
reasonable to base a business decision on such a small difference.
Enter the idea of Effect Size. The name is descriptive but at the same time not very
illuminating on what this measure does. We will get to specific measures shortly, but for now,
let’s look at how an Effect Size measure can help us understand our findings. First, the name:
Effect Size. What effect? What size? In very general terms, the effect we are monitoring is the
effect that occurs when we change one of the variables. For example, is there an effect on the
average compa-ratio when we change from male to female. Certainly, but not all that much, as
we found no significant difference between the average male and female compa-ratios. Is there
an effect when we change from male to female on the average salary? Definitely. And it is
much larger than what we observed on the compa-ratio means. We found a significant
difference in the average salary for males than females – around $14,000.
The Effect Siz.
1) The path length from A to B in the following graph is .docxmonicafrancis71118
1) The path length from A to B in the following graph is:
a- 2
b- 10
c- 22
d- There is no path
2) The minimum path weight from A to B in the following graph is:
a- 2
b- 10
c- 32
d- There is no path
3) The minimum path weight from A to E in the following graph is:
a- 1
b- 7
c- 67
d- There is no path
4) The longest cycle that starts at A and ends at A in the following graph is:
a- 104
b- 122
c- 42
d- There is no cycle
5) The entry AE in the length one adjacency matrix representation of the following graph is:
a- 7
b-
c- 0
d- None of the above
6) The entry AB in the length one adjacency matrix representation of the following graph is:
a- 10
b-
c- 22
d- 0
7) The entry AD in the length two adjacency matrix representation of the following graph is:
a- 60
b-
c- 44
d- 0
8) In the following graph, which of the following paths is considered a simple path?
a- AECAD
b- AEBFC
c- ADBFD
d- There is no simple path in the graph above
9) Some of the cliques the following graph has include: (A clique is a subgraph that is complete which means each node in the subgraph is connected to every other node n the subgraph). In the following graph, the subgraph AEBD is not a clique because A and B are not connected and E and D are not connected also, otherwise if they were connected it would be a clique.
a- ADBE, EBFC, EB, F, C
b- AEC, DBF
c- AEB, EBC
d- AECFBD
10) (TSP): Apply the nearest-neighbor algorithm to the complete weighted graph G in the following figure, beginning at vertex B, what is the path and the total weight?
a- BADECB with weight 725
b- BAEDCB with weight 775
c- TSP does not work with complete graph
d- None of the answers is true
Experimental Design 1
Running Head: EXPERIMENTAL DESIGN
Experimental Design and Some Threats to
Experimental Validity: A Primer
Susan Skidmore
Texas A&M University
Paper presented at the annual meeting of the Southwest Educational
Research Association, New Orleans, Louisiana, February 6, 2008.
Experimental Design 2
Abstract
Experimental designs are distinguished as the best method to respond to
questions involving causality. The purpose of the present paper is to explicate
the logic of experimental design and why it is so vital to questions that demand
causal conclusions. In addition, types of internal and external validity threats are
discussed. To emphasize the current interest in experimental designs, Evidence-
Based Practices (EBP) in medicine, psychology and education are highlighted.
Finally, cautionary statements regarding experimental designs are elucidated
with examples from the literature.
Experimental Design 3
The No Child Left Behind Act (NCLB) demands “scientifically based
research” as the basis for awarding many grants in education (2001).
Specifically, the 107th Congress (2001) delineated scientifically-based research
as that which “is evaluated using experimen.
Using Cloud-based statistics applications to enhance statistics educationsmackinnon
Slides to accompany my 2019 presentation at the CPA. Discusses my approach to teaching statistics using online applications and active learning workshops.
These are slides I use when teaching my second year undergraduate statistics course. They are designed more for conceptual understanding, and do not have syntax for programs like SPSS or R. So it is a more conceptual and mathematical review, rather than a "how-to" computer guide.
Generalized Linear Models for Between-Subjects Designssmackinnon
Here is a tutorial on how to use generalized linear models in SPSS software. These are models that are frequently more appropriate than ANOVA or linear regression, especially when the distributions of outcome variables are non-normal and/or homogeneity of variance assumptions are violated.
Increasing Power without Increasing Sample Sizesmackinnon
This is an invited presentation I gave at a symposium "Making your research more reproducible" at the 27th Annual Conference of the Association for Psychological Science, New York. It talks about increasing statistical power without increasing sample size.
These are some slides I use in my Multivariate Statistics course to teach psychology graduate student the basics of structural equation modeling using the lavaan package in R. Topics are at an introductory level, for someone without prior experience with the topic.
A gentle introduction to growth curves using SPSSsmackinnon
A brief introduction on how to conduct growth curve statistical analyses using SPSS software, including some sample syntax. Originally presented at IWK Statistics Seminar Series at the IWK Health Center, Halifax, NS, May 1, 2013.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
1. Mediation in health research:
A statistics workshop using SPSS
Dr. Sean P. Mackinnon
Dalhousie University
Crossroads Interdisciplinary Health Conference, 2015
2. What kinds of questions does
mediation answer?
• Mediation asks about the process by which a
predictor variable affects an outcome
• “Does X predict M, which in turn predicts Y?”
• E.g., “Does exercise improve cardiovascular
health, which in turn increases longevity?”
3. Linear Regression
• Understanding mediation requires a basic
understanding of linear regression
• Displayed as a path diagram, it could look
something like this:
Impulsivity Binge Drinking
.30
The number depicted here is the slope (B value, or b1 above)
c-path
also called the “total effect”
iii XbbY 10
4. Mediation
• Mediation builds on this basic linear regression model by
adding a third variable (i.e., the “mediator”)
• In mediation, the third variable is thought to come in
between X & Y. So, X leads to the mediator, which in turn
leads to Y.
Impulsivity Binge Drinking
Enhancement
Motives
5. Mediation
• The idea is, the c-path (the direct effect) should get smaller
with the addition of a mediator.
• So, we want to know if the c-path – c’-path is “statistically
significant.”
Impulsivity Binge Drinking
Enhancement
Motives
c’-path
Also called the “direct effect”
6. Mediation
• To test this, you first need to get the slope of two other
relationships: a and b paths
Impulsivity Binge Drinking
Enhancement
Motives
c’-path
Get the slope of this
relationship
a-path
Get the slope of this
relationship while also
controlling for
enhancement motives
b-path
7. Mediation
• Mathematicians have shown that
– (a-path * b-path) = c-path – c’ path
– (But only when X and M are continuous)
• Thus, if a*b (“the indirect effect”) is statistically significant,
mediation has occurred
Impulsivity Binge Drinking
Enhancement
Motives
c’-path
a-path b-path
Preacher & Hayes (2008)
8. Significance of Indirect Effect
• Lots of ways to test the significance of a*b
– Test of Joint Significance
– Sobel Test
– Bootstrapped Confidence Intervals
• Of these methods, bootstrapping is currently the most preferred
• But … Hayes & Scharkow (2013) have shown that the different
methods agree > 90% of the time…
9. Joint Significance Test
(Baron & Kenny, 1986)
• If the a-path AND the b-path are both significant,
conclude that a*b is also significant.
• This is a liberal test (i.e., high Type I error) and is
usually used as a supplement to other methods.
Impulsivity Binge Drinking
Enhancement
Motives
.05
.25* .28*
c’ path
a-path b-path
10. Sobel Test (Sobel, 1982)
• An alternative is to estimate the indirect effect and its significance
using the Sobel test (Sobel. 1982).
• It is a conservative test (i.e., high Type II error)
• z-value = a*b/SQRT(b2*sa
2 + a2*sb
2)
– a = B value (slope) for a-path
– b = B value (slope) for b-path
– sa = SE for a-path
– sa = SE for b-path
• Online Calculator for Sobel Test:
– http://quantpsy.org/sobel/sobel.htm
– Also available in the PROCESS macro discussed later
11. Bootstrapping
• The sobel test is inaccurate because it relies on an
assumption of a normal sampling distrbution:
– However, the sampling distribution distribution of a*b is
non-normal except in very large samples…
• Bootstrapping is a computer intensive, robust analysis
technique that can be applied to non-normal data.
• Virtually any analysis can be bootstrapped, but we’re
going to apply it to testing the significance of the
indirect effect (a*b).
12. What is a “Re-Sample?”
In SPSS, Each row is a “person” who has an ID, and lots of values on measures
A “re-sample” randomly samples participants from the sample, with replacement
Re-sample 1
ID1
ID3
ID4
ID2
Re-sample 2
ID1
ID1
ID3
ID2
Re-sample 3
ID4
ID4
ID2
ID2
Note that people can be duplicated in the resamples using this method
13. What is bootstrapping?
The idea of the sampling distribution of the sample mean x-bar: take
very many samples, collect the x-values from each, and look at the
distribution of these values
From Hesterberg et al. (2003)
14. What is bootstrapping?
From Hesterberg et al. (2003)
The theory shortcut: if we know that the population values follow
a normal distribution, theory tells us that the sampling
distribution of x-bar is also normal.
This is known as the
central limit theorem
15. What is bootstrapping?
From Hesterberg et al. (2003)
The bootstrap idea: when theory fails and we can afford only one
sample, that sample stands in for the population, and
the distribution of x in many resamples stands in for the sampling
distribution
16. Bootstrapping Indirect Effects
• Create 1000s of simulated datasets using re-
sampling with replacement
– Pretends as though your sample is the population, and
you simulate other samples from that.
• Run the analysis once in each of these 1000s of
samples
• Of those analyses, 95% of the generated statistics
will fall between two numbers. If zero isn’t in that
interval, p < .05!
17. Effect Sizes for Mediation
• There are many different ways to calculate effect
sizes for mediation analysis (Preacher & Kelly, 2011)
• Two simple-to-understand effect size measures are:
– Percent mediation (PM)
– Completely Standardized Indirect Effect (abcs)
18. Percent Mediation
Impulsivity Binge Drinking
Enhancement
Motives
.12* (.05)
.25* .28*
c-path (c’ path)
a-path b-path
ab = .25 * .28 = .07
c = .12
PM = .07 / .12 = .583
Interpreted as the percent of the total effect (c) accounted
for by your indirect effect (a*b).
19. Note about Percent Mediation…
• The direct effect (c’-path) can sometimes be
larger than the total effect (c-path)
– Inconsistent mediation
• In these cases, take the absolute value of c’
before calculating effect size to avoid
proportions greater than 1.0.
20. Completely Standardized Indirect
Effect
• So, it’s just two steps:
– 1. Calculate the standardized regression paths for the a and b
paths
– 2. Multiply them together to get the ES
– (So, just standardize your variables before analysis and you can
get a 95% CI!)
• Is now a standardized version that will be similar in
interpretation across measures … but it’s no longer
bounded by -1 and 1 like a correlation.
Which is the
same as …
21. Installing the PROCESS macro in SPSS
• Download files from here:
– process.spd
– http://www.processmacro.org/download.html
Once you do this, you’ll get a new analysis
you can run under:
Analyze Regression PROCESS
Now every time you open SPSS, you’ll
have the option to run mediation analyses!
22. A Sample Model w. Output
Conscientious
Personality
Overall Physical
Health
Health-Related
Behaviours
Uses a (fabricated) dataset you can find online here if
you want to try it on your own time for practice:
http://savvystatistics.com/wp-
content/uploads/2015/03/crossroads.2015.data_.csv
RQ: Do health related behaviours mediate the relationship between
conscientious personality and overall physical health?
23. How to Run in SPSS
For basic mediation, use “model 4”
Conscientiousness = X
Physical health = Y
Health-Related Behaviours = M
24. Annotated Output: a, b. c’ paths
Coeff = Slope; SE = standard error; t = t-statistic; p = p-value
LLCI & ULCI = lower and upper levels for confidence interval
a-path
b-path
c'-path (direct effect)
26. Annotated Output: Effect Size &
Significance of Indirect Effect
Effect Size 1: abcs
(Report the 95% CI For this)
Effect Size 2: PM
(Don’t use the 95% CI For this)
Upper and Lower
Bootstrapped 95% CI
a*b or “indirect effect”
Report the 95% CI for this
If the CI for a*b does not include
zero, then mediation has occurred!
27. Reporting Mediation Analysis
There was a significant indirect effect of
conscientiousness on overall physical health through
health-related behaviours, ab = 0.21, BCa CI [0.15,
0.26]. The mediator could account for roughly half of
the total effect, PM = .44.
Conscientious
Personality
Overall Physical
Health
Health-Related
Behaviours0.52*** 0.39***
0.26***
(0.47)***
29. Appendix: Syntax
*Make sure to run the process.sps macro first, or
this won’t work!
*This is an alternative to running using the GUI
PROCESS vars = health bfi.c behave
/y=health/x=bfi.c/m=behave/w=/z=/v=/q=/
model =4/boot=1000/center=0/hc3=1/effsize=1/
normal=1/coeffci=1/conf=95/percent=0/total=1/
covmy=0/jn=0/quantile =0/plot=0/contrast=0/
decimals=F10.4/covcoeff=0.
2015-03-24