Testing Mediation and regression analysisOsama Yousaf
Mediation occurs when a third variable (the mediator) intervenes between a predictor and an outcome. Baron and Kenny proposed a four-step approach using regression to test for mediation: 1) predict the outcome from the predictor, 2) predict the mediator from the predictor, 3) predict the outcome from the mediator, 4) predict the outcome from both the predictor and mediator. Full mediation is supported if the predictor is no longer significant when controlling for the mediator, while partial mediation occurs if both remain significant. The indirect effect represents the mediated portion and can be estimated by the difference or product of coefficients; significance is tested using methods like Sobel's test or bootstrapping.
This document provides information on calculating effect sizes when comparing two means. It defines effect size as the extent to which a phenomenon is present in a population or how false the null hypothesis is. It lists several common effect size measures for different statistical tests, including Cohen's d for independent groups t-tests, correlation coefficients for correlational analyses, and eta squared and omega squared for ANOVA. An example is given of computing Cohen's d to compare study habits between public and private school students using t-test results.
This document discusses important concepts for screening data, including detecting and handling errors, missing data, outliers, and ensuring assumptions of analyses are met. It describes why data screening is important to obtain accurate results and avoid bias. Key topics covered include identifying patterns of missing data, different types of missing data (MCAR, MAR, MNAR), and various methods for treating missing values. Outliers are defined and their impact explained. Common transformations are presented to achieve normality, linearity, and homoscedasticity. Checklists are provided for conducting data screening.
This presentation discusses the procedure involved in two-way mixed ANOVA design. The procedure has been discussed by solving a problem using SPSS functionality.
1) The document presents a statistical modeling approach called targeted smooth Bayesian causal forests (tsbcf) to smoothly estimate heterogeneous treatment effects over gestational age using observational data from early medical abortion regimens.
2) The tsbcf method extends Bayesian additive regression trees (BART) to estimate treatment effects that evolve smoothly over gestational age, while allowing for heterogeneous effects across patient subgroups.
3) The tsbcf analysis of early medical abortion regimen data found the simultaneous administration to be similarly effective overall to the interval administration, but identified some patient subgroups where effectiveness may vary more over gestational age.
1. The document discusses generalized linear mixed models (GLMMs), which are statistical models that combine linear predictors, non-normal response distributions, link functions, and random effects.
2. It outlines some of the statistical, computational, and sociological challenges in using GLMMs, such as estimating models with large matrices and interpreting results accurately.
3. The conclusion emphasizes next steps like improving correlation structures and inference methods in GLMMs while addressing issues like proper interpretation and use by non-experts.
This document discusses statistical methods for comparing two independent sample means and two independent sample proportions. It provides steps and examples for conducting significance tests to compare population means and proportions. For means, it describes using a z-test where the test statistic is the difference between sample means divided by the pooled standard error. For proportions, it describes using a z-test where the test statistic is the difference between sample proportions divided by the pooled standard error. Examples provided show conducting these tests to analyze differences in housework hours and attitudes between years.
We describe different approaches for specifying models and prior distributions for estimating heterogeneous treatment effects using Bayesian nonparametric models. We make an affirmative case for direct, informative (or partially informative) prior distributions on heterogeneous treatment effects, especially when treatment effect size and treatment effect variation is small relative to other sources of variability. We also consider how to provide scientifically meaningful summaries of complicated, high-dimensional posterior distributions over heterogeneous treatment effects with appropriate measures of uncertainty.
Testing Mediation and regression analysisOsama Yousaf
Mediation occurs when a third variable (the mediator) intervenes between a predictor and an outcome. Baron and Kenny proposed a four-step approach using regression to test for mediation: 1) predict the outcome from the predictor, 2) predict the mediator from the predictor, 3) predict the outcome from the mediator, 4) predict the outcome from both the predictor and mediator. Full mediation is supported if the predictor is no longer significant when controlling for the mediator, while partial mediation occurs if both remain significant. The indirect effect represents the mediated portion and can be estimated by the difference or product of coefficients; significance is tested using methods like Sobel's test or bootstrapping.
This document provides information on calculating effect sizes when comparing two means. It defines effect size as the extent to which a phenomenon is present in a population or how false the null hypothesis is. It lists several common effect size measures for different statistical tests, including Cohen's d for independent groups t-tests, correlation coefficients for correlational analyses, and eta squared and omega squared for ANOVA. An example is given of computing Cohen's d to compare study habits between public and private school students using t-test results.
This document discusses important concepts for screening data, including detecting and handling errors, missing data, outliers, and ensuring assumptions of analyses are met. It describes why data screening is important to obtain accurate results and avoid bias. Key topics covered include identifying patterns of missing data, different types of missing data (MCAR, MAR, MNAR), and various methods for treating missing values. Outliers are defined and their impact explained. Common transformations are presented to achieve normality, linearity, and homoscedasticity. Checklists are provided for conducting data screening.
This presentation discusses the procedure involved in two-way mixed ANOVA design. The procedure has been discussed by solving a problem using SPSS functionality.
1) The document presents a statistical modeling approach called targeted smooth Bayesian causal forests (tsbcf) to smoothly estimate heterogeneous treatment effects over gestational age using observational data from early medical abortion regimens.
2) The tsbcf method extends Bayesian additive regression trees (BART) to estimate treatment effects that evolve smoothly over gestational age, while allowing for heterogeneous effects across patient subgroups.
3) The tsbcf analysis of early medical abortion regimen data found the simultaneous administration to be similarly effective overall to the interval administration, but identified some patient subgroups where effectiveness may vary more over gestational age.
1. The document discusses generalized linear mixed models (GLMMs), which are statistical models that combine linear predictors, non-normal response distributions, link functions, and random effects.
2. It outlines some of the statistical, computational, and sociological challenges in using GLMMs, such as estimating models with large matrices and interpreting results accurately.
3. The conclusion emphasizes next steps like improving correlation structures and inference methods in GLMMs while addressing issues like proper interpretation and use by non-experts.
This document discusses statistical methods for comparing two independent sample means and two independent sample proportions. It provides steps and examples for conducting significance tests to compare population means and proportions. For means, it describes using a z-test where the test statistic is the difference between sample means divided by the pooled standard error. For proportions, it describes using a z-test where the test statistic is the difference between sample proportions divided by the pooled standard error. Examples provided show conducting these tests to analyze differences in housework hours and attitudes between years.
We describe different approaches for specifying models and prior distributions for estimating heterogeneous treatment effects using Bayesian nonparametric models. We make an affirmative case for direct, informative (or partially informative) prior distributions on heterogeneous treatment effects, especially when treatment effect size and treatment effect variation is small relative to other sources of variability. We also consider how to provide scientifically meaningful summaries of complicated, high-dimensional posterior distributions over heterogeneous treatment effects with appropriate measures of uncertainty.
This document discusses generalized linear mixed models (GLMMs). It begins with examples of GLMM applications and definitions of key terms. The document then covers estimation methods for GLMMs, including maximum likelihood estimation, integrated likelihood, and both deterministic and stochastic approaches. Inference for GLMMs and remaining challenges are also mentioned. The overall document provides an overview of GLMM frameworks, examples, estimation techniques, and open questions.
I. The median test is used to determine if two independent groups have been drawn from populations with the same median. It requires at least ordinal scale data.
II. The combined median of both groups is calculated. Scores from each group are then split based on whether they are above or below the combined median. These frequencies are entered into a 2x2 contingency table.
III. The median test statistic (chi-square) is calculated and compared to a critical value based on the significance level and degrees of freedom to determine whether to reject or fail to reject the null hypothesis that the two groups have the same median.
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...IOSRJM
-This research work investigated the behaviour of a new semiparametric non-linear (SPNL) model on
a set of panel data with very small time point (T = 1). The SPNL model incorporates the relationship between
individual independent variable and unobserved heterogeneity variable. Five different estimation techniques
namely; Least Square (LS), Generalized Method of Moments (GMM), Continuously Updating (CU), Empirical
Likelihood (EL) and Exponential Tilting (ET) Estimators were employed for the estimation; for the purpose of
modelling the metrical response variable non-linearly on a set of independent variables. The performances of
these estimators on the SPNL model were examined for different parameters in the model using the Least
Square Error (LSE), Mean Absolute Error (MAE) and Median Absolute Error (MedAE) criteria at the lowest
time point (T = 1). The results showed that the ET estimator which provided the least errors of estimation is
relatively more efficient for the proposed model than any of the other estimators considered. It is therefore
recommended that the ET estimator should be employed to estimate the SPNL model for panel data with very
small time point.
This study examined the effects of social anxiety disorder (SAD) on reward and punishment learning in 80 veterans with unipolar depression. Participants completed two signal detection tasks to assess responses to receiving rewards and punishments. Results showed no difference in reward learning between those with depression alone versus depression and SAD. However, individuals with both depression and SAD showed increased sensitivity to punishment compared to those with depression alone, performing better at avoiding punishments. This suggests SAD contributes additively to increased punishment-based learning among individuals with co-occurring depression and SAD. The findings have implications for developing therapeutic strategies focused on reducing avoidance of punishment feedback in this comorbid group.
Group 3 analyzed data set 39 to examine relationships between self-esteem, education, and age. For research question 1, ANOVA found no significant difference in self-esteem levels between education groups. For research question 2, an independent t-test found that older age groups had significantly higher self-esteem than younger groups. The report included sample descriptions, hypothesis testing, statistical analyses and conclusions for both research questions.
Structural equation modeling (SEM) is used to analyze relationships between multiple independent and dependent variables. It allows for simultaneous testing of these relationships while accounting for measurement error. The goal of SEM is to determine if the estimated population covariance matrix from the model fits the sample covariance matrix. It can be used to test theories, account for variance, and assess reliability and parameter estimates. Key considerations include sample size, normality, linearity, and identification of the model. Model fit is assessed using absolute, comparative, and parsimonious fit indices. Modification indices can also indicate how to improve model fit.
The document provides an overview of two-factor ANOVA, including:
- Two-factor ANOVA involves more than one independent variable (IV) and evaluates three main hypotheses - the main effects of each IV and their interaction.
- It partitions the total variance into between-treatments variance and within-treatments variance. Between-treatments variance is further partitioned into portions attributable to each IV and their interaction.
- F-ratios are calculated to test the three hypotheses by comparing the between-treatments mean squares to the within-treatments mean squares. If an F-ratio exceeds the critical value, its hypothesis is supported.
This document provides an overview of generalized linear mixed models (GLMMs). It begins with examples and definitions, then discusses estimation methods like maximum likelihood estimation. It describes how random effects are used to account for correlation in grouped data. Estimation balances fitting fixed effects to the data and fitting random effects to their assumed distribution. The document outlines inference challenges and open questions with GLMMs. It indicates Wald tests are commonly used but can provide poor approximations in some cases.
Stability criterion of periodic oscillations in a (4)Alexander Decker
1) The authors establish that the distribution of the harmonic mean of group variances is a generalized beta distribution through simulation.
2) They show that the generalized beta distribution can be approximated by a chi-square distribution.
3) This means that the harmonic mean of group variances is approximately chi-square distributed, though the degrees of freedom need not be an integer. Using the harmonic mean in place of the pooled variance allows hypothesis testing when group variances are unequal.
Week 5 Lecture 14 The Chi Square TestQuite often, patterns of .docxcockekeshia
Week 5 Lecture 14
The Chi Square Test
Quite often, patterns of responses or measures give us a lot of information. Patterns are generally the result of counting how many things fit into a particular category. Whenever we make a histogram, bar, or pie chart we are looking at the pattern of the data. Frequently, changes in these visual patterns will be our first clues that things have changed, and the first clue that we need to initiate a research study (Lind, Marchel, & Wathen, 2008).
One of the most useful test in examining patterns and relationships in data involving counts (how many fit into this category, how many into that, etc.) is the chi-square. It is extremely easy to calculate and has many more uses than we will cover. Examining patterns involves two uses of the Chi-square - the goodness of fit and the contingency table. Both of these uses have a common trait: they involve counts per group. In fact, the chi-square is the only statistic we will look at that we use when we have counts per multiple groups (Tanner & Youssef-Morgan, 2013). Chi Square Goodness of Fit Test
The goodness of fit test checks to see if the data distribution (counts per group) matches some pattern we are interested in. Example: Are the employees in our example company distributed equal across the grades? Or, a more reasonable expectation for a company might be are the employees distributed in a pyramid fashion – most on the bottom and few at the top?
The Chi Square test compares the actual versus a proposed distribution of counts by generating a measure for each cell or count: (actual – expected)2/actual. Summing these for all of the cells or groups provides us with the Chi Square Statistic. As with our other tests, we determine the p-value of getting a result as large or larger to determine if we reject or not reject our null hypothesis. An example will show the approach using Excel.
Regardless of the Chi Square test, the chi square related functions are found in the fx Statistics window rather than the Data Analysis where we found the t and ANOVA test functions. The most important for us are:
· CHISQ.TEST (actual range, expected range) – returns the p-value for the test
· CHISQ.INV.RT(p-value, df) – returns the actual Chi Square value for the p-value or probability value used.
· CHISQ.DIST.RT(X, df) – returns the p-value for a given value.
When we have a table of actual and expected results, using the =CHISQ.TEST(actual range, expected range) will provide us with the p-value of the calculated chi square value (but does not give us the actual calculated chi square value for the test). We can compare this value against our alpha criteria (generally 0.05) to make our decision about rejecting or not rejecting the null hypothesis.
If, after finding the p-value for our chi square test, we want to determine the calculated value of the chi square statistic, we can use the =CHISQ.INV.RT(probability, df) function, the value for probability is .
This document provides an overview and introduction to an econometrics course. It discusses how econometrics can be used to estimate quantitative causal effects by using data and observational studies. Examples discussed include estimating the effect of class size on student achievement. The document outlines how the course will cover methods for estimating causal effects using observational data, with a focus on applications. It also reviews key probability and statistics concepts needed for the course, including probability distributions, moments, hypothesis testing, and the sampling distribution. The document presents an example analysis using data on class sizes and test scores to illustrate initial estimation, hypothesis testing, and confidence interval techniques.
The document summarizes a simulation study that examined the effects of using raw scores versus IRT-derived scores when operationalizing latent constructs in moderated multiple regression analyses. The study found that using raw scores can inflate Type 1 error rates for interaction terms under conditions of assessment inappropriateness. However, rescaling the scores using the Graded Response Model, a polytomous IRT model, mitigated these effects. The study supports the idea that IRT scores provide a more robust metric than raw scores in moderated regression analyses, especially under suboptimal assessment conditions.
Investigations of certain estimators for modeling panel data under violations...Alexander Decker
This document investigates the efficiency of four methods for estimating panel data models (pooling, first differencing, between, and feasible generalized least squares) when the assumptions of homoscedasticity, no autocorrelation, and no collinearity are jointly violated. Monte Carlo simulations were conducted under varying conditions of heteroscedasticity, autocorrelation, collinearity, sample size, and time periods. The results showed that in small samples, the feasible generalized least squares estimator is most efficient when heteroscedasticity is severe, regardless of autocorrelation and collinearity levels. However, when heteroscedasticity is low to moderate with moderate autocorrelation, first differencing and feasible generalized least squares
This document summarizes a study that used the fuzzy TOPSIS method to select the optimal type of spillway for a dam in northern Greece called Pigi Dam. Five alternative spillway types were evaluated based on nine criteria. The criteria were expressed as triangular fuzzy numbers to account for uncertainty. Weights for the criteria were determined using the AHP method and also expressed linguistically as fuzzy numbers. The fuzzy TOPSIS method was then used to rank the alternatives based on their distances from the ideal and negative-ideal solutions. The alternative with the highest relative closeness to the ideal solution was determined to be the optimal spillway type.
No support for declining effect sizes over time - Chris C Martin and Gregory ...Chris Martin
This document presents the findings of three meta-meta-analyses that examined evidence for the decline effect over time. Study 1 analyzed 3,488 effect sizes from 70 meta-analytic tables and found no significant correlation between effect size and year of publication. Study 2 analyzed 37 social psychology articles and found that 62.2% reported flat trends over time. Study 3 analyzed 33 clinical psychology articles and found that 80% reported flat trends over time. Overall, the studies found no strong evidence that effect sizes consistently decline with increasing replications.
This document summarizes a study that assessed the stability of 20 wheat genotypes grown in 40 environments in Pakistan using nonparametric methods. The data exhibited severe heterogeneity and violated assumptions of normality and homogeneity of variances required for parametric analyses. Nonparametric stability methods were applied that are robust to these assumption violations. The modified rank-sum method identified genotypes G7, G3, G15, G5 and G12 as most stable and high yielding, while G14 and G19 were least stable. Nonparametric methods provided a justified alternative for analyzing genotype-environment interactions in this heteroscedastic and non-normal data.
Model of robust regression with parametric and nonparametric methodsAlexander Decker
This document summarizes and compares several parametric and nonparametric methods for estimating the parameters in a simple linear regression model when outliers are present in the data. It introduces ordinary least squares regression as the classical parametric method and discusses its limitations when outliers are present. It then summarizes several nonparametric and robust regression methods that are less influenced by outliers, including Theil's method, least absolute deviations regression, M-estimation, and trimmed least squares regression. The document presents the models and algorithms for these various methods. It concludes by describing a simulation study that evaluates and compares the performance of these different estimation techniques under various types and amounts of outliers.
Week 5 Lecture 14 The Chi Square Test Quite often, pat.docxcockekeshia
Week 5 Lecture 14
The Chi Square Test
Quite often, patterns of responses or measures give us a lot of information. Patterns are
generally the result of counting how many things fit into a particular category. Whenever we
make a histogram, bar, or pie chart we are looking at the pattern of the data. Frequently, changes
in these visual patterns will be our first clues that things have changed, and the first clue that we
need to initiate a research study (Lind, Marchel, & Wathen, 2008).
One of the most useful test in examining patterns and relationships in data involving
counts (how many fit into this category, how many into that, etc.) is the chi-square. It is
extremely easy to calculate and has many more uses than we will cover. Examining patterns
involves two uses of the Chi-square - the goodness of fit and the contingency table. Both of
these uses have a common trait: they involve counts per group. In fact, the chi-square is the only
statistic we will look at that we use when we have counts per multiple groups (Tanner &
Youssef-Morgan, 2013).
Chi Square Goodness of Fit Test
The goodness of fit test checks to see if the data distribution (counts per group) matches
some pattern we are interested in. Example: Are the employees in our example company
distributed equal across the grades? Or, a more reasonable expectation for a company might be
are the employees distributed in a pyramid fashion – most on the bottom and few at the top?
The Chi Square test compares the actual versus a proposed distribution of counts by
generating a measure for each cell or count: (actual – expected)2/actual. Summing these for all
of the cells or groups provides us with the Chi Square Statistic. As with our other tests, we
determine the p-value of getting a result as large or larger to determine if we reject or not reject
our null hypothesis. An example will show the approach using Excel.
Regardless of the Chi Square test, the chi square related functions are found in the fx
Statistics window rather than the Data Analysis where we found the t and ANOVA test
functions. The most important for us are:
• CHISQ.TEST (actual range, expected range) – returns the p-value for the test
• CHISQ.INV.RT(p-value, df) – returns the actual Chi Square value for the p-value
or probability value used.
• CHISQ.DIST.RT(X, df) – returns the p-value for a given value.
When we have a table of actual and expected results, using the =CHISQ.TEST(actual
range, expected range) will provide us with the p-value of the calculated chi square value (but
does not give us the actual calculated chi square value for the test). We can compare this value
against our alpha criteria (generally 0.05) to make our decision about rejecting or not rejecting
the null hypothesis.
If, after finding the p-value for our chi square test, we want to determine the calculated
value of the chi square statistic, we can use the =CHISQ.INV.RT(probability, df).
Invited talk at the Focus Fortnight 8: ""The analysis of discrete choice experiments", organized by the Centre for Bayesian Statistics in Health Economics, University of Sheffield (UK), September, 2007.
This document summarizes methods for subgroup identification in clinical trials. It begins by distinguishing predictive from prognostic biomarkers. It then provides a taxonomy of four main approaches to subgroup identification: global outcome modeling, global treatment effect modeling, modeling individual treatment regimes, and local treatment effect modeling (subgroup search). The document discusses several examples and methods under each approach. It concludes by noting important considerations for evaluating subgroup identification methods, such as the number of predictors handled, model complexity control, type I error control, and obtaining honest effect size estimates.
This document summarizes a discussion between Susan Athey and Guido Imbens on the relationship between machine learning and causal inference. It notes that while machine learning excels at prediction problems using large datasets, it has weaknesses when it comes to causal questions. Econometrics and statistics literature focuses more on formal theories of causality. The document proposes combining the strengths of both fields by developing machine learning methods that can estimate causal effects, accounting for issues like endogeneity and treatment effect heterogeneity. It outlines some open problems and directions for future research at the intersection of these fields.
This document discusses generalized linear mixed models (GLMMs). It begins with examples of GLMM applications and definitions of key terms. The document then covers estimation methods for GLMMs, including maximum likelihood estimation, integrated likelihood, and both deterministic and stochastic approaches. Inference for GLMMs and remaining challenges are also mentioned. The overall document provides an overview of GLMM frameworks, examples, estimation techniques, and open questions.
I. The median test is used to determine if two independent groups have been drawn from populations with the same median. It requires at least ordinal scale data.
II. The combined median of both groups is calculated. Scores from each group are then split based on whether they are above or below the combined median. These frequencies are entered into a 2x2 contingency table.
III. The median test statistic (chi-square) is calculated and compared to a critical value based on the significance level and degrees of freedom to determine whether to reject or fail to reject the null hypothesis that the two groups have the same median.
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...IOSRJM
-This research work investigated the behaviour of a new semiparametric non-linear (SPNL) model on
a set of panel data with very small time point (T = 1). The SPNL model incorporates the relationship between
individual independent variable and unobserved heterogeneity variable. Five different estimation techniques
namely; Least Square (LS), Generalized Method of Moments (GMM), Continuously Updating (CU), Empirical
Likelihood (EL) and Exponential Tilting (ET) Estimators were employed for the estimation; for the purpose of
modelling the metrical response variable non-linearly on a set of independent variables. The performances of
these estimators on the SPNL model were examined for different parameters in the model using the Least
Square Error (LSE), Mean Absolute Error (MAE) and Median Absolute Error (MedAE) criteria at the lowest
time point (T = 1). The results showed that the ET estimator which provided the least errors of estimation is
relatively more efficient for the proposed model than any of the other estimators considered. It is therefore
recommended that the ET estimator should be employed to estimate the SPNL model for panel data with very
small time point.
This study examined the effects of social anxiety disorder (SAD) on reward and punishment learning in 80 veterans with unipolar depression. Participants completed two signal detection tasks to assess responses to receiving rewards and punishments. Results showed no difference in reward learning between those with depression alone versus depression and SAD. However, individuals with both depression and SAD showed increased sensitivity to punishment compared to those with depression alone, performing better at avoiding punishments. This suggests SAD contributes additively to increased punishment-based learning among individuals with co-occurring depression and SAD. The findings have implications for developing therapeutic strategies focused on reducing avoidance of punishment feedback in this comorbid group.
Group 3 analyzed data set 39 to examine relationships between self-esteem, education, and age. For research question 1, ANOVA found no significant difference in self-esteem levels between education groups. For research question 2, an independent t-test found that older age groups had significantly higher self-esteem than younger groups. The report included sample descriptions, hypothesis testing, statistical analyses and conclusions for both research questions.
Structural equation modeling (SEM) is used to analyze relationships between multiple independent and dependent variables. It allows for simultaneous testing of these relationships while accounting for measurement error. The goal of SEM is to determine if the estimated population covariance matrix from the model fits the sample covariance matrix. It can be used to test theories, account for variance, and assess reliability and parameter estimates. Key considerations include sample size, normality, linearity, and identification of the model. Model fit is assessed using absolute, comparative, and parsimonious fit indices. Modification indices can also indicate how to improve model fit.
The document provides an overview of two-factor ANOVA, including:
- Two-factor ANOVA involves more than one independent variable (IV) and evaluates three main hypotheses - the main effects of each IV and their interaction.
- It partitions the total variance into between-treatments variance and within-treatments variance. Between-treatments variance is further partitioned into portions attributable to each IV and their interaction.
- F-ratios are calculated to test the three hypotheses by comparing the between-treatments mean squares to the within-treatments mean squares. If an F-ratio exceeds the critical value, its hypothesis is supported.
This document provides an overview of generalized linear mixed models (GLMMs). It begins with examples and definitions, then discusses estimation methods like maximum likelihood estimation. It describes how random effects are used to account for correlation in grouped data. Estimation balances fitting fixed effects to the data and fitting random effects to their assumed distribution. The document outlines inference challenges and open questions with GLMMs. It indicates Wald tests are commonly used but can provide poor approximations in some cases.
Stability criterion of periodic oscillations in a (4)Alexander Decker
1) The authors establish that the distribution of the harmonic mean of group variances is a generalized beta distribution through simulation.
2) They show that the generalized beta distribution can be approximated by a chi-square distribution.
3) This means that the harmonic mean of group variances is approximately chi-square distributed, though the degrees of freedom need not be an integer. Using the harmonic mean in place of the pooled variance allows hypothesis testing when group variances are unequal.
Week 5 Lecture 14 The Chi Square TestQuite often, patterns of .docxcockekeshia
Week 5 Lecture 14
The Chi Square Test
Quite often, patterns of responses or measures give us a lot of information. Patterns are generally the result of counting how many things fit into a particular category. Whenever we make a histogram, bar, or pie chart we are looking at the pattern of the data. Frequently, changes in these visual patterns will be our first clues that things have changed, and the first clue that we need to initiate a research study (Lind, Marchel, & Wathen, 2008).
One of the most useful test in examining patterns and relationships in data involving counts (how many fit into this category, how many into that, etc.) is the chi-square. It is extremely easy to calculate and has many more uses than we will cover. Examining patterns involves two uses of the Chi-square - the goodness of fit and the contingency table. Both of these uses have a common trait: they involve counts per group. In fact, the chi-square is the only statistic we will look at that we use when we have counts per multiple groups (Tanner & Youssef-Morgan, 2013). Chi Square Goodness of Fit Test
The goodness of fit test checks to see if the data distribution (counts per group) matches some pattern we are interested in. Example: Are the employees in our example company distributed equal across the grades? Or, a more reasonable expectation for a company might be are the employees distributed in a pyramid fashion – most on the bottom and few at the top?
The Chi Square test compares the actual versus a proposed distribution of counts by generating a measure for each cell or count: (actual – expected)2/actual. Summing these for all of the cells or groups provides us with the Chi Square Statistic. As with our other tests, we determine the p-value of getting a result as large or larger to determine if we reject or not reject our null hypothesis. An example will show the approach using Excel.
Regardless of the Chi Square test, the chi square related functions are found in the fx Statistics window rather than the Data Analysis where we found the t and ANOVA test functions. The most important for us are:
· CHISQ.TEST (actual range, expected range) – returns the p-value for the test
· CHISQ.INV.RT(p-value, df) – returns the actual Chi Square value for the p-value or probability value used.
· CHISQ.DIST.RT(X, df) – returns the p-value for a given value.
When we have a table of actual and expected results, using the =CHISQ.TEST(actual range, expected range) will provide us with the p-value of the calculated chi square value (but does not give us the actual calculated chi square value for the test). We can compare this value against our alpha criteria (generally 0.05) to make our decision about rejecting or not rejecting the null hypothesis.
If, after finding the p-value for our chi square test, we want to determine the calculated value of the chi square statistic, we can use the =CHISQ.INV.RT(probability, df) function, the value for probability is .
This document provides an overview and introduction to an econometrics course. It discusses how econometrics can be used to estimate quantitative causal effects by using data and observational studies. Examples discussed include estimating the effect of class size on student achievement. The document outlines how the course will cover methods for estimating causal effects using observational data, with a focus on applications. It also reviews key probability and statistics concepts needed for the course, including probability distributions, moments, hypothesis testing, and the sampling distribution. The document presents an example analysis using data on class sizes and test scores to illustrate initial estimation, hypothesis testing, and confidence interval techniques.
The document summarizes a simulation study that examined the effects of using raw scores versus IRT-derived scores when operationalizing latent constructs in moderated multiple regression analyses. The study found that using raw scores can inflate Type 1 error rates for interaction terms under conditions of assessment inappropriateness. However, rescaling the scores using the Graded Response Model, a polytomous IRT model, mitigated these effects. The study supports the idea that IRT scores provide a more robust metric than raw scores in moderated regression analyses, especially under suboptimal assessment conditions.
Investigations of certain estimators for modeling panel data under violations...Alexander Decker
This document investigates the efficiency of four methods for estimating panel data models (pooling, first differencing, between, and feasible generalized least squares) when the assumptions of homoscedasticity, no autocorrelation, and no collinearity are jointly violated. Monte Carlo simulations were conducted under varying conditions of heteroscedasticity, autocorrelation, collinearity, sample size, and time periods. The results showed that in small samples, the feasible generalized least squares estimator is most efficient when heteroscedasticity is severe, regardless of autocorrelation and collinearity levels. However, when heteroscedasticity is low to moderate with moderate autocorrelation, first differencing and feasible generalized least squares
This document summarizes a study that used the fuzzy TOPSIS method to select the optimal type of spillway for a dam in northern Greece called Pigi Dam. Five alternative spillway types were evaluated based on nine criteria. The criteria were expressed as triangular fuzzy numbers to account for uncertainty. Weights for the criteria were determined using the AHP method and also expressed linguistically as fuzzy numbers. The fuzzy TOPSIS method was then used to rank the alternatives based on their distances from the ideal and negative-ideal solutions. The alternative with the highest relative closeness to the ideal solution was determined to be the optimal spillway type.
No support for declining effect sizes over time - Chris C Martin and Gregory ...Chris Martin
This document presents the findings of three meta-meta-analyses that examined evidence for the decline effect over time. Study 1 analyzed 3,488 effect sizes from 70 meta-analytic tables and found no significant correlation between effect size and year of publication. Study 2 analyzed 37 social psychology articles and found that 62.2% reported flat trends over time. Study 3 analyzed 33 clinical psychology articles and found that 80% reported flat trends over time. Overall, the studies found no strong evidence that effect sizes consistently decline with increasing replications.
This document summarizes a study that assessed the stability of 20 wheat genotypes grown in 40 environments in Pakistan using nonparametric methods. The data exhibited severe heterogeneity and violated assumptions of normality and homogeneity of variances required for parametric analyses. Nonparametric stability methods were applied that are robust to these assumption violations. The modified rank-sum method identified genotypes G7, G3, G15, G5 and G12 as most stable and high yielding, while G14 and G19 were least stable. Nonparametric methods provided a justified alternative for analyzing genotype-environment interactions in this heteroscedastic and non-normal data.
Model of robust regression with parametric and nonparametric methodsAlexander Decker
This document summarizes and compares several parametric and nonparametric methods for estimating the parameters in a simple linear regression model when outliers are present in the data. It introduces ordinary least squares regression as the classical parametric method and discusses its limitations when outliers are present. It then summarizes several nonparametric and robust regression methods that are less influenced by outliers, including Theil's method, least absolute deviations regression, M-estimation, and trimmed least squares regression. The document presents the models and algorithms for these various methods. It concludes by describing a simulation study that evaluates and compares the performance of these different estimation techniques under various types and amounts of outliers.
Week 5 Lecture 14 The Chi Square Test Quite often, pat.docxcockekeshia
Week 5 Lecture 14
The Chi Square Test
Quite often, patterns of responses or measures give us a lot of information. Patterns are
generally the result of counting how many things fit into a particular category. Whenever we
make a histogram, bar, or pie chart we are looking at the pattern of the data. Frequently, changes
in these visual patterns will be our first clues that things have changed, and the first clue that we
need to initiate a research study (Lind, Marchel, & Wathen, 2008).
One of the most useful test in examining patterns and relationships in data involving
counts (how many fit into this category, how many into that, etc.) is the chi-square. It is
extremely easy to calculate and has many more uses than we will cover. Examining patterns
involves two uses of the Chi-square - the goodness of fit and the contingency table. Both of
these uses have a common trait: they involve counts per group. In fact, the chi-square is the only
statistic we will look at that we use when we have counts per multiple groups (Tanner &
Youssef-Morgan, 2013).
Chi Square Goodness of Fit Test
The goodness of fit test checks to see if the data distribution (counts per group) matches
some pattern we are interested in. Example: Are the employees in our example company
distributed equal across the grades? Or, a more reasonable expectation for a company might be
are the employees distributed in a pyramid fashion – most on the bottom and few at the top?
The Chi Square test compares the actual versus a proposed distribution of counts by
generating a measure for each cell or count: (actual – expected)2/actual. Summing these for all
of the cells or groups provides us with the Chi Square Statistic. As with our other tests, we
determine the p-value of getting a result as large or larger to determine if we reject or not reject
our null hypothesis. An example will show the approach using Excel.
Regardless of the Chi Square test, the chi square related functions are found in the fx
Statistics window rather than the Data Analysis where we found the t and ANOVA test
functions. The most important for us are:
• CHISQ.TEST (actual range, expected range) – returns the p-value for the test
• CHISQ.INV.RT(p-value, df) – returns the actual Chi Square value for the p-value
or probability value used.
• CHISQ.DIST.RT(X, df) – returns the p-value for a given value.
When we have a table of actual and expected results, using the =CHISQ.TEST(actual
range, expected range) will provide us with the p-value of the calculated chi square value (but
does not give us the actual calculated chi square value for the test). We can compare this value
against our alpha criteria (generally 0.05) to make our decision about rejecting or not rejecting
the null hypothesis.
If, after finding the p-value for our chi square test, we want to determine the calculated
value of the chi square statistic, we can use the =CHISQ.INV.RT(probability, df).
Invited talk at the Focus Fortnight 8: ""The analysis of discrete choice experiments", organized by the Centre for Bayesian Statistics in Health Economics, University of Sheffield (UK), September, 2007.
This document summarizes methods for subgroup identification in clinical trials. It begins by distinguishing predictive from prognostic biomarkers. It then provides a taxonomy of four main approaches to subgroup identification: global outcome modeling, global treatment effect modeling, modeling individual treatment regimes, and local treatment effect modeling (subgroup search). The document discusses several examples and methods under each approach. It concludes by noting important considerations for evaluating subgroup identification methods, such as the number of predictors handled, model complexity control, type I error control, and obtaining honest effect size estimates.
This document summarizes a discussion between Susan Athey and Guido Imbens on the relationship between machine learning and causal inference. It notes that while machine learning excels at prediction problems using large datasets, it has weaknesses when it comes to causal questions. Econometrics and statistics literature focuses more on formal theories of causality. The document proposes combining the strengths of both fields by developing machine learning methods that can estimate causal effects, accounting for issues like endogeneity and treatment effect heterogeneity. It outlines some open problems and directions for future research at the intersection of these fields.
ABSTRACT : This paper critically examined a broad view of Structural Equation Model (SEM) with a view
of pointing out direction on how researchers can employ this model to future researches, with specific focus on
several traditional multivariate procedures like factor analysis, discriminant analysis, path analysis. This study
employed a descriptive survey and historical research design. Data was computed viaDescriptive Statistics,
Correlation Coefficient, Reliability. The study concluded that Novice researchers must take care of assumptions
and concepts of Structure Equation Modeling, while building a model to check the proposed hypothesis. SEM is
more or less an evolving technique in the research, which is expanding to new fields. Moreover, it is providing
new insights to researchers for conducting longitudinal investigations.
.
These are some slides I use in my Multivariate Statistics course to teach psychology graduate student the basics of structural equation modeling using the lavaan package in R. Topics are at an introductory level, for someone without prior experience with the topic.
Statistical modelling is of prime importance in each and every sphere of data analysis. This paper reviews the justification of fitting linear model to the collected data. Inappropriateness of the fitted model may be due two reasons 1.wrong choice of the analytical form, 2. Suffers from the adverse effects of outliers and/or influential observations. The aim is to identify outliers using the deletion technique. In I extend the result of deletion diagnostics to the ex- changeable model and reviews some results of model analytical form checking and the technique illustrated through an example.
RESIDUALS AND INFLUENCE IN NONLINEAR REGRESSION FOR REPEATED MEASUREMENT DATAorajjournal
All observations don’t have equal significance in regression analysis. Diagnostics of observations is an important aspect of model building. In this paper, we use diagnostics method to detect residuals and influential points in nonlinear regression for repeated measurement data. Cook distance and Gauss newton method have been proposed to identify the outliers in nonlinear regression analysis and parameter estimation. Most of these techniques based on graphical representations of residuals, hat matrix and case deletion measures. The results
show us detection of single and multiple outliers cases in repeated measurement data. We use these techniques
to explore performance of residuals and influence in nonlinear regression model.
This document discusses the lack of emphasis on research design in quantitative psychology methods education. It summarizes two surveys of PhD programs from 1990 and 2008 that found gains in statistics training but no improvement in research design instruction. While competence in laboratory experiments was high, it was weak for other designs like field experiments and longitudinal studies. The document also reviews Donald Campbell's influential work developing concepts of internal and external validity threats and remedies. It notes the slow diffusion of newer approaches to causal inference from statistics, like propensity scores, into psychology. Overall it argues for revitalizing research design education with more application examples and engaging younger faculty across fields.
Similar to Correcting for unreliability & partial invariance: A two-stage path analysis approach (20)
• For a full set of 530+ questions. Go to
https://skillcertpro.com/product/servicenow-cis-itsm-exam-questions/
• SkillCertPro offers detailed explanations to each question which helps to understand the concepts better.
• It is recommended to score above 85% in SkillCertPro exams before attempting a real exam.
• SkillCertPro updates exam questions every 2 weeks.
• You will get life time access and life time free updates
• SkillCertPro assures 100% pass guarantee in first attempt.
1.) Introduction
Our Movement is not new; it is the same as it was for Freedom, Justice, and Equality since we were labeled as slaves. However, this movement at its core must entail economics.
2.) Historical Context
This is the same movement because none of the previous movements, such as boycotts, were ever completed. For some, maybe, but for the most part, it’s just a place to keep your stable until you’re ready to assimilate them into your system. The rest of the crabs are left in the world’s worst parts, begging for scraps.
3.) Economic Empowerment
Our Movement aims to show that it is indeed possible for the less fortunate to establish their economic system. Everyone else – Caucasian, Asian, Mexican, Israeli, Jews, etc. – has their systems, and they all set up and usurp money from the less fortunate. So, the less fortunate buy from every one of them, yet none of them buy from the less fortunate. Moreover, the less fortunate really don’t have anything to sell.
4.) Collaboration with Organizations
Our Movement will demonstrate how organizations such as the National Association for the Advancement of Colored People, National Urban League, Black Lives Matter, and others can assist in creating a much more indestructible Black Wall Street.
5.) Vision for the Future
Our Movement will not settle for less than those who came before us and stopped before the rights were equal. The economy, jobs, healthcare, education, housing, incarceration – everything is unfair, and what isn’t is rigged for the less fortunate to fail, as evidenced in society.
6.) Call to Action
Our movement has started and implemented everything needed for the advancement of the economic system. There are positions for only those who understand the importance of this movement, as failure to address it will continue the degradation of the people deemed less fortunate.
No, this isn’t Noah’s Ark, nor am I a Prophet. I’m just a man who wrote a couple of books, created a magnificent website: http://www.thearkproject.llc, and who truly hopes to try and initiate a truly sustainable economic system for deprived people. We may not all have the same beliefs, but if our methods are tried, tested, and proven, we can come together and help others. My website: http://www.thearkproject.llc is very informative and considerably controversial. Please check it out, and if you are afraid, leave immediately; it’s no place for cowards. The last Prophet said: “Whoever among you sees an evil action, then let him change it with his hand [by taking action]; if he cannot, then with his tongue [by speaking out]; and if he cannot, then, with his heart – and that is the weakest of faith.” [Sahih Muslim] If we all, or even some of us, did this, there would be significant change. We are able to witness it on small and grand scales, for example, from climate control to business partnerships. I encourage, invite, and challenge you all to support me by visiting my website.
11June 2024. An online pre-engagement session was organized on Tuesday June 11 to introduce the Science Policy Lab approach and the main components of the conceptual framework.
About 40 experts from around the globe gathered online for a pre-engagement session, paving the way for the first SASi-SPi Science Policy Lab event scheduled for June 18-19, 2024 in Malmö. The session presented the objectives for the upcoming Science Policy Lab (S-PoL), which featured a role-playing game designed to simulate stakeholder interactions and policy interventions for food systems transitions. Participants called for the sharing of meeting materials and continued collaboration, reflecting a strong commitment to advancing towards sustainable agrifood systems.
Kalyan chart satta matka guessing resultsanammadhu484
MATKASATTABOSS.COM IS INDIA'S MOST TRUSTED NO.1 WEBSITE. WE PROVIDE YOU EXACT GUESSING OF THE MATKA RESULT BY OUR TOP GUESSER, MATKASATTABOSS.COM ALWAYS PROVIDES EXACT AND FAST MATKA RESULTS. PLAY SATTA MATKA AND BECOME SATTA KING BY THE HELP OF MATKASATTABOSS.COM. INDIA'S TOP SATTA MATKA MARKET AND THEIR FAST MATKA RESULTS. GET ALL THE RESULTS AND WIN MONEY BY PERFECT KALYAN MATKA TIPS , MATKA GUESSING BY OUR TOP GUESSER AND KALYAN RAJSHREE RAJYOG SWASTIK NATRAAJ BANGLORE BIRLA RAJDHANI MILAN TIME BAZAAR MATKA CHART .
Gamify it until you make it Improving Agile Development and Operations with ...Ben Linders
So many challenges, so little time. While we’re busy developing software and keeping it operational, we also need to sharpen the saw, but how? Gamification can be a way to look at how you’re doing and find out where to improve. It’s a great way to have everyone involved and get the best out of people.
In this presentation, Ben Linders will show how playing games with the DevOps coaching cards can help to explore your current development and deployment (DevOps) practices and decide as a team what to improve or experiment with.
The games that we play are based on an engagement model. Instead of imposing change, the games enable people to pull in ideas for change and apply those in a way that best suits their collective needs.
By playing games, you can learn from each other. Teams can use games, exercises, and coaching cards to discuss values, principles, and practices, and share their experiences and learnings.
Different game formats can be used to share experiences on DevOps principles and practices and explore how they can be applied effectively. This presentation provides an overview of playing formats and will inspire you to come up with your own formats.
6. 6
Joint Modeling Not Always
Practical
Need a large model
E.g., 3 structural coefficients, but ~ 100
parameters with JM
Sample size implication
8. 8
Joint Modeling Not Always
Practical
Computational challenges with discrete indicators
Maximum likelihood
numerical integration with high dimensions
Weighted least squares
may need N >= 200
missing data handling
10. 10
2S-PA
First stage: Obtain one indicator ( ) per latent
construct ( )
E.g., regression scores; EAP scores
Adjust for noninvariance
Second stage: path modeling with and standard
error of measurement/reliability
Available in standard psychometric software
Lai & Hsiao* (2021, Psychological Methods); Lai, Tse*, Zhang*, Li*, & Hsiao*
(under review)
19. 19
Summary
Separate estimation seems a good alternative
strategy for adjust both noninvariance and
unreliability
Plus better small sample performance
Lai, Tse*, Zhang*, Li*, & Hsiao* (under review)
20. 20
References
Cole, D. A., & Preacher, K. J. (2014). Manifest variable path analysis:
Potentially serious and misleading consequences due to uncorrected
measurement error. Psychological Methods, 19(2), 300–315.
https://doi.org/10.1037/a0033805
Hsiao, Y.-Y., & Lai, M. H. C. (2018). The impact of partial
measurement invariance on testing moderation for single and multi-
level data. Frontiers in Psychology, 9, Article 740.
https://doi.org/10.3389/fpsyg.2018.00740
Lai, M. H. C., & Hsiao, Y.-Y. (2021). Two-stage path analysis with
definition variables: An alternative framework to account for
measurement error. Psychological Methods.
https://doi.org/10.1037/met0000410
Lai, Tse*, Zhang*, Li*, & Hsiao* (under review)
Good morning, thank you for having me. It’s been a great learning experience for me hearing from the previous presentations. Today I'm excited to share with you a recent work on two-stage path analysis. Specifically, I’ll talk about how to use this approach to correct for both unreliability and partial invariance. This is joint work with my grad students, Winnie Tse & Gengrui Zhang, Yixiao Li who’s an undergrad in my lab, & my colleague Yu-Yu Hsiao
As we know, measures in the social and behavioral sciences are usually not 100% accurate, and there could be systematic differences due to different individual and contextual factors, which we call violations of measurement invariance, or simply noninvariance
For example, the graph here shows that for the same true depression level, females, the red line, tend to get a higher observed score than males. Therefore, any gender differences we found on the test scores could be just due to noninvariance.
As a result, noninvariance can lead to biased or spurious group differences, and can also lead to biased or spurious interactions.
Consider the graph here with two groups, where the regression line is the same for the two groups if we have an accurate measurement of Y. When noninvariance exists, it can push the scores for group 1 upward, and the association between the test and the latent variable may change.
The effect of measurement noninvariance is usually accompanied by the presence of unreliability, because our measures are rarely, if ever, perfectly reliable. It is well known in the literature that unreliability can bias regression slopes.
But when the reliability is different across groups, it creates differential impact, as shown on the right, where the reliability is lower for G2, in blue than for G1, in red.
Therefore, we can get very misleading results if we don’t pay attention to measurement bias and measurement unreliability.
A usual solution is to do what we call joint modeling, where we model both the relation between the indicators and the latent variables, which is the measurement model, and also the relations among the latent variables, or the structural model, simultaneously
However, as many of you who have experience with SEM modeling may know, joint modeling is not always practical. For one thing, joint modeling requires using all indicators, and the number of indicators is usually much larger than the number of constructs. So for the graph in the last page, even though there is only three constructs, we need to estimate like 100 parameters.
As such, we may need a large sample size for stable estimation
And in small samples, the more complex the model, the more likely we run into convergence issues
In addition, there are computational challenges when we have discrete items or indicators. In the frequentist framework, one usually either uses maximum likelihood or weighted least squares estimations.
The challenge with ML is that it requires numerical integration, and so is not feasible in high dimensional problems.
With WLS, it generally requires even a larger sample size, and have a stricter requirement on missing data.
As an alternative, we can consider a two-stage estimation strategy, using what we call two-stage path analysis, abbreviated as 2S-PA. We first proposed this in a recent paper, currently in press in psychological methods
As opposed to joint modeling, the 2S-PA approach only requires one indicator per construct, so the structural model is much easier to specify. And by reducing the model size, it also has a smaller sample size requirement.
In the first stage of 2S-PA, we use psychometric analyses to obtain the best estimates of the latent variables. We call these estimates eta tilde. For example, we can obtain different kinds of factor scores with factor analysis or item response theory. It is also important in this stage that we account for measurement noninvariance, like using a partial invariance model.
In the second stage, we treat eta tilde as single indicators of the latent variables in the path model. In addition, we also need the standard error of measurement, or the reliability, of eta tilde. These are usually available in standard psychometric software. So with these, we can account for partial invariance in the first stage, and the unreliability in the second stage.
Here’s a bit more detail. In the first stage, say we have a construct, eta, of interest. For each construct, we perform a psychometric analysis to obtain an estimate or prediction, which we call eta titlde, of each person’s latent variable score. For many models, eta tilde can be expressed as a times eta plus error, where a is a scaling constant, and the error follows classical measurement error with mean 0, and is uncorrelated with eta.
As an example, the first column here shows the factor scores for a latent variable fm, the second column is the standard error of measurement, and the third is the reliability of the factor scores.
When the indicators are continuous, the standard error of measurement is constant, but for discrete indicators, the standard error is usually person-specific, like the numbers in the last two columns.
After we obtain an indicator for each construct in the first stage, the second stage is basically a structural equation model with single indicators.
However, notice the subscript i here, which allows the standard error of measurement to be different across observations. This is needed when the indicators were obtained using models like IRT with discrete indicators.
With single indicators, the model is identified by constraining the measurement error variance based on the standard error of measurement values in the first stage.
The indicator loading is set to 1 for convenience, so the latent variables are generally scaled differently than in the first stage. However, this is not a problem if we’re interested in the standardized coefficients, meaning we rescale the parameters so that the latent variables have variances of 1.
While the idea of two stage estimation is not new, previous methods usually do not consider both noninvariance and unreliability. Also, an advantage of 2S-PA is it can handle not just continuous indicators, but also discrete items.
With discrete items, the error variance is usually different for different individuals, but this can be handled using definition variables. The idea is, instead of setting a common constraint for every observation, we can allow the constraint to depend on a variable, like the one represented in diamond in the diagram, for individual-specific error variance.
The estimation can be done in standard software for structural equation modeling, such as Mplus and OpenMx, with maximum likelihood.
It can also be done with Bayesian estimation such as in Stan.
To compare 2S-PA and joint modeling, we conducted a simulation study. The paper is currently under review. Actually we just got the invitation to revise a few days ago. We simulated data from a mediation model with three variables, the one shown in an earlier slide. The treatment variable, X, is observed, while the mediator, eta M, is measured by six continuous indicators, and the outcome, eta Y, is measured by 16 binary indicators.
The simulated data have two groups, with sample size conditions of 50, 100, and 300 per group.
There were three noninvariant items for eta M, and five noninvariant items for eta Y, across the two groups.
Here’s a summary of the results. First we look at the convergence rate. As you can see in the table, joint modeling has a lot of convergence issues, like when n is 50 per group, the convergence rate was only 10%. On the other hand, with 2S-PA it was 92%. While convergence rates improved when the sample size increased, even with 300 observations, joint modeling was still only at about 80% convergence rate, whereas 2S-PA achieved close to 100% convergence rate with just 100 observations per group.
The graph here shows the parameter bias. The methods are shown in the x-axis. FS-PA means factor score path analysis without adjusting for measurement error, which, as expected, gave biased estimates when the true coefficients were not zero. This is the classic attenuation due to measurement error.
For JM and 2S-PA, we can see they were close to unbiased when the sample sizes are large, but in smaller samples, 2S-PA outperformed JM.
We also look at inferences. Here I only compare 2S-PA and JM, as FS-PA does not give consistent estimates. The x-axis is the sample size, and the y-axis is the coverage of the 95% CI. We want the empirical coverage rates to be close to or above 95%. As you can see, with JM, which is represented by the blue lines, the coverage of the 95% confidence intervals was quite low in small samples.
On the other hand, with 2S-PA, it performed quite well even in small samples. This includes all the path coefficients, as well as the product term beta_1 beta_3, or the indirect effect.
To summarize the study, we found that the separate estimation strategy in 2S-PA seems a good alternative for adjusting both noninvariance and unreliability, especially in small samples.
I want to end by thanking Oi-man for organizing this great opportunity, and allowing me to be part of it. Thank you, you audience, for staying with me. I’m looking forward to your questions and suggestions.