2 nd Order CFA Byrne Chapter 5. 2 nd Order Models The idea of a 2 nd order model (sometimes called a bi-factor model) is: – You have some latent variables.

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

MCMC Estimation MCMC = Markov chain Monte Carlo
Structural Equation Modeling. What is SEM Swiss Army Knife of Statistics Can replicate virtually any model from “canned” stats packages (some limitations.
Structural Equation Modeling
Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
Confidence Intervals This chapter presents the beginning of inferential statistics. We introduce methods for estimating values of these important population.
Correlation and regression Dr. Ghada Abo-Zaid
Theoretical Probability Distributions We have talked about the idea of frequency distributions as a way to see what is happening with our data. We have.
Random variable Distribution. 200 trials where I flipped the coin 50 times and counted heads no_of_heads in a trial.
Terminology A statistic is a number calculated from a sample of data. For each different sample, the value of the statistic is a uniquely determined number.
Bayesian statistics – MCMC techniques
1 Analysis of Variance This technique is designed to test the null hypothesis that three or more group means are equal.
Point estimation, interval estimation
1 More Regression Information. 2 3 On the previous slide I have an Excel regression output. The example is the pizza sales we saw before. The first thing.
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
The Simple Regression Model
Statistical Analysis SC504/HS927 Spring Term 2008 Week 17 (25th January 2008): Analysing data.
5-3 Inference on the Means of Two Populations, Variances Unknown
Sampling Distributions & Point Estimation. Questions What is a sampling distribution? What is the standard error? What is the principle of maximum likelihood?
Getting Started with Hypothesis Testing The Single Sample.
Standard Error of the Mean
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
Hypothesis Testing:.
Confidence Intervals and Hypothesis Testing
SECTION 6.4 Confidence Intervals for Variance and Standard Deviation Larson/Farber 4th ed 1.
Inference for Regression
Priors, Normal Models, Computing Posteriors
Chapter 8: Confidence Intervals
Stat 13, Tue 5/8/ Collect HW Central limit theorem. 3. CLT for 0-1 events. 4. Examples. 5.  versus  /√n. 6. Assumptions. Read ch. 5 and 6.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
SEM: Basics Byrne Chapter 1 Tabachnick SEM
Path Analysis. Remember What Multiple Regression Tells Us How each individual IV is related to the DV The total amount of variance explained in a DV Multiple.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Estimation Kline Chapter 7 (skip , appendices)
A Comparison of Two MCMC Algorithms for Hierarchical Mixture Models Russell Almond Florida State University College of Education Educational Psychology.
Confirmatory Factor Analysis Psych 818 DeShon. Construct Validity: MTMM ● Assessed via convergent and divergent evidence ● Convergent – Measures of the.
Full Structural Models Kline Chapter 10 Brown Chapter 5 ( )
1 Chapter 8 Sampling Distributions of a Sample Mean Section 2.
Section 10.1 Confidence Intervals
2nd Order CFA Hierarchical Latent Models
Multigroup Models Byrne Chapter 7 Brown Chapter 7.
Confidence Intervals Lecture 3. Confidence Intervals for the Population Mean (or percentage) For studies with large samples, “approximately 95% of the.
Multigroup Models Beaujean Chapter 4 Brown Chapter 7.
CFA: Basics Beaujean Chapter 3. Other readings Kline 9 – a good reference, but lumps this entire section into one chapter.
Chapter 10: Confidence Intervals
SEM: Basics Byrne Chapter 1 Tabachnick SEM
MTMM Byrne Chapter 10. MTMM Multi-trait multi-method – Traits – the latent factors you are trying to measure – Methods – the way you are measuring the.
Assumptions 5.4 Data Screening. Assumptions Parametric tests based on the normal distribution assume: – Independence – Additivity and linearity – Normality.
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
Sampling and estimation Petter Mostad
Chapter 13.  Both Principle components analysis (PCA) and Exploratory factor analysis (EFA) are used to understand the underlying patterns in the data.
ANOVA, Regression and Multiple Regression March
Chapter 13 Sampling distributions
Estimation Kline Chapter 7 (skip , appendices)
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
CFA Model Revision Byrne Chapter 4 Brown Chapter 5.
Lecture 8: Ordinary Least Squares Estimation BUEC 333 Summer 2009 Simon Woodcock.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Chapter 9 Minitab Recipe Cards. Contingency tests Enter the data from Example 9.1 in C1, C2 and C3.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
Tests of Significance We use test to determine whether a “prediction” is “true” or “false”. More precisely, a test of significance gets at the question.
Chapter 17 STRUCTURAL EQUATION MODELING. Structural Equation Modeling (SEM)  Relatively new statistical technique used to test theoretical or causal.
Data Screening. What is it? Data screening is very important to make sure you’ve met all your assumptions, outliers, and error problems. Each type of.
CFA: Basics Byrne Chapter 3 Brown Chapter 3 (40-53)
Bootstrapping Byrne Chapter 12. Bootstrapping Quick distinction: – Population distribution – Sample distribution – Sampling distribution.
Markov Chain Monte Carlo in R
CHAPTER 8 Estimating with Confidence
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
CHAPTER 29: Multiple Regression*
Presentation transcript:

2 nd Order CFA Byrne Chapter 5

2 nd Order Models The idea of a 2 nd order model (sometimes called a bi-factor model) is: – You have some latent variables that are measured by the observed variables – A portion of the variance in the latent variables can be explained by a second (set) of latent variables

2 nd Order Models Therefore, we are switching out the covariances between factors and using another latent to explain them.

Identification Remember that each portion of the model has to be identified. – The section with each latent variable has to be identified (so you need at least one loading set to 1). – The section with the latents has to be identified (draw)

Identification You can do get over identification in a couple of ways: – Set some of the loadings in the upper portion of the model to be equal (give them the same name) – You can set the variance in the upper latent to be 1 – You can set some of the error variances of the latents in the lower portion to be equal

Critical Ratio of Differences Lists the CR of the difference between parameter loadings – Remember that CR are basically Z scores Located in the analysis properties -> output window

Critical Ratio of Differences When you get this output, all the parameters will get labels (to be able to tell what’s going on) The chart will look like a correlation table (similar to residual moments) – You are looking for parameters with very small values close to zero. – That means they have basically no difference in magnitude

Critical Ratio of Differences The logic: – If two parameters are basically the same, why are we wasting a degree of freedom estimating the second one?

Let’s Try It!

A quick review SEM assumes that the latent variables are continuous – Well, sometimes that isn’t the data we actually have.

Issues Things that can happen when you assume continuous but they aren’t: – Correlations appear lower than they are, especially for items with low #s of categories – Chi-square values are inflated, especially when items are both positive and negative skewed

Issues Things that can happen when you assume continuous but they aren’t: – Loadings are underestimated – SEs are underestimated

Categorical Data But! If you have 4-5+ categories and they approximate normal = you are probably ok Solutions: – Special correlation tables + ADF estimation – Bayesian estimation!

Bayesian Estimation Definitions: – Prior distribution: a guess at what the underlying distribution of the parameter might be (this sounds crazy! How am I supposed to know! But think about the fact that we traditionally assume distributions are normal for analyses, so is it so crazy?) You just have to guess! But Amos does this for you actually.

Bayesian Estimation Definitions: – Posterior distribution: distribution of the parameter after you have analyzed the data + the influence of the prior distribution So you use the data to estimate the parameters but include a little bit of the prior distribution as part of your estimate for the parameter

Bayesian Estimation Example of prior and posterior Two things to notice: – How much data I have matters: the posterior is less influenced by the prior when you have more data – How strong I make my prior matters: the posterior is less influenced when you make a weak prior

Bayesian Estimation How to in Amos: – Go to Analysis properties (seriously, everything minus SRMR is in that window) – Turn on estimate means and intercepts If you forget this step, you will get an error message saying “EEK!”

Bayesian Estimation To run Bayes, click on the button with the little distribution on it OR Analyze > Bayesian Estimation

So what’s going on?

Things to Note MCMC what? – Markov Chain Monte Carlo – Sometimes called a random walk procedure – Example

Things to Note The pause button – This button stops the analysis from running ## – 500 is called the BURN IN – Basically all MCMC analyses require a little bit of time before they settle down. – I like to think of it as a drunken walk at the beginning so you exclude that part.

Things to Note ### – The number part is how many steps the program has run to converge (come to a stable solution). – It’s going to be a big number, as many Markov Chains have to run 50,000 times to get a stable solution.

Things to Note How do I know I have a stable solution? – Unhappy face – Happy face and below result in happy faces But 1.10 is a common criteria as well You can stop the algorithm at any time

Things to Note Next you get the loadings, means, intercepts, covariances, etc. The MEAN column is your estimate SE = standard error for all those samples – will be small SD = the estimate of SE for ML estimation CS is the convergence statistic

Things to Note Posterior Icon – Gives you the posterior distribution – You can get the first + last distribution You want these to overlap a lot

Things to Note Autocorrelation – This fancy word is the problem of the starting point. – If values are correlated, it implies that you didn’t get a good convergence of the data … aka when the walk started you got stuck somewhere or distracted (not a good thing).

Things to Note You want the autocorrelation picture to be positively skewed and drop off to zero at the end.

Things to Note Trace plots – Trace plots show you the walk the MCMC chain took. It should look like a very messy bunch of squiggles – A problem (usually autocorrelation) would be if it looked like a bunch of lines together, then a break, then a bunch of lines (draw)

Let’s Try it! Let’s run a Bayesian analysis on Byrne’s second order model (you need full data for Bayes, not just correlations)