Assessing Hypothesis Testing Fit Indices Kline Chapter 8 (Stop at 210) Byrne page 68-84.

Slides:



Advertisements
Similar presentations
Structural Equation Modeling
Advertisements

Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Statistical Decision Making
Testing Hypotheses About Proportions Chapter 20. Hypotheses Hypotheses are working models that we adopt temporarily. Our starting hypothesis is called.
SOC 681 James G. Anderson, PhD
Chapter 10 Simple Regression.
Multivariate Data Analysis Chapter 11 - Structural Equation Modeling.
The Simple Regression Model
Chapter 11 Multiple Regression.
Chapter 3 Hypothesis Testing. Curriculum Object Specified the problem based the form of hypothesis Student can arrange for hypothesis step Analyze a problem.
G Lecture 51 Estimation details Testing Fit Fit indices Arguing for models.
1 Inference About a Population Variance Sometimes we are interested in making inference about the variability of processes. Examples: –Investors use variance.
1 The Sample Mean rule Recall we learned a variable could have a normal distribution? This was useful because then we could say approximately.
Today Concepts underlying inferential statistics
Multivariate Analysis of Variance, Part 1 BMTRY 726.
Scot Exec Course Nov/Dec 04 Ambitious title? Confidence intervals, design effects and significance tests for surveys. How to calculate sample numbers when.
Stages in Structural Equation Modeling
How Can We Test whether Categorical Variables are Independent?
Confirmatory factor analysis
AM Recitation 2/10/11.
Chapter 13: Inference in Regression
Structural Equation Modeling 3 Psy 524 Andrew Ainsworth.
CENTRE FOR INNOVATION, RESEARCH AND COMPETENCE IN THE LEARNING ECONOMY Session 2: Basic techniques for innovation data analysis. Part I: Statistical inferences.
One-Factor Experiments Andy Wang CIS 5930 Computer Systems Performance Analysis.
1 Tests with two+ groups We have examined tests of means for a single group, and for a difference if we have a matched sample (as in husbands and wives)
Confirmatory Factor Analysis Psych 818 DeShon. Purpose ● Takes factor analysis a few steps further. ● Impose theoretically interesting constraints on.
Today’s lesson Confidence intervals for the expected value of a random variable. Determining the sample size needed to have a specified probability of.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Comparing two sample means Dr David Field. Comparing two samples Researchers often begin with a hypothesis that two sample means will be different from.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Six.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
1 General Structural Equations (LISREL) Week 1 #4.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Multigroup Models Byrne Chapter 7 Brown Chapter 7.
Assessing Hypothesis Testing Fit Indices
Measures of Fit David A. Kenny January 25, Background Introduction to Measures of Fit.
CFA: Basics Beaujean Chapter 3. Other readings Kline 9 – a good reference, but lumps this entire section into one chapter.
Chapter 3: Statistical Significance Testing Warner (2007). Applied statistics: From bivariate through multivariate. Sage Publications, Inc.
SEM: Basics Byrne Chapter 1 Tabachnick SEM
MTMM Byrne Chapter 10. MTMM Multi-trait multi-method – Traits – the latent factors you are trying to measure – Methods – the way you are measuring the.
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
Exploratory Factor Analysis Principal Component Analysis Chapter 17.
SEM Model Fit: Introduction David A. Kenny January 12, 2014.
Examples. Path Model 1 Simple mediation model. Much of the influence of Family Background (SES) is indirect.
ALISON BOWLING CONFIRMATORY FACTOR ANALYSIS. REVIEW OF EFA Exploratory Factor Analysis (EFA) Explores the data All measured variables are related to every.
The general structural equation model with latent variates Hans Baumgartner Penn State University.
Uncertainty and confidence Although the sample mean,, is a unique number for any particular sample, if you pick a different sample you will probably get.
Chapter 7: Hypothesis Testing. Learning Objectives Describe the process of hypothesis testing Correctly state hypotheses Distinguish between one-tailed.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Class Seven Turn In: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 For Class Eight: Chapter 20: 18, 20, 24 Chapter 22: 34, 36 Read Chapters 23 &
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
The SweSAT Vocabulary (word): understanding of words and concepts. Data Sufficiency (ds): numerical reasoning ability. Reading Comprehension (read): Swedish.
CFA: Basics Byrne Chapter 3 Brown Chapter 3 (40-53)
Bootstrapping Byrne Chapter 12. Bootstrapping Quick distinction: – Population distribution – Sample distribution – Sampling distribution.
Measures of Fit David A. Kenny.
More on Inference.
Structural Equation Modeling using MPlus
Chapter 4. Inference about Process Quality
CJT 765: Structural Equation Modeling
More on Inference.
Writing about Structural Equation Models
Factors Affecting Measures of Fit
Confirmatory Factor Analysis
Factors Affecting Measures of Fit
Testing Causal Hypotheses
Presentation transcript:

Assessing Hypothesis Testing Fit Indices Kline Chapter 8 (Stop at 210) Byrne page 68-84

Fit Indices It’s complicated yo. – There are a lot of them. – They do not imply that you are perfectly right. – They have guidelines but they are not perfect either. – People misuse them. – Etc.

Fit Indices Limitations: – Fit statistics indicate an overall/average model fit. That means there can be bad sections, but the overall fit is good. – No one magical number/summary. – They do not tell you where a misspecification occurs.

Fit Indices Limitations – Do not tell you the predictive value of a model. – Do not tell you if it’s theoretically meaningful.

Fit Indices Model test statistic – examines if the reproduced correlation matrix matches the sample correlation matrix – Sometimes called “badness of fit” – Want these to be small

Fit Indices Traditional NHST = reject-support context – You reject the null to show your research hypothesis is correct. SEM Hyp Testing = accept-support context – You do not reject the null showing that your model is consistent with the normal

Fit Indices Both types of statistical inferences have their problems … and especially in SEM because it is easy to find statistics that you would normally reject, even with good model fit. Tends to be too black and white (reject or not to reject!)

Fit Indices Approximate Fit Indices – Not traditionally a dichotomous yes-no decision – Do not distinguish between sampling error and evidence against the model

Fit Indices Approximate Fit Indices – Absolute Fit Indices – Incremental Fit Indices – A parsimony-adjusted Index – Predictive Fit Indices

Fit Indices Absolute fit indices – Proportion of the covariance matrix explained by the model – You can think about these as sort of R 2 – Want these values be high

Fit Indices Incremental Fit Indices – Also known as comparative fit indices – Compared to the improvement over the independence model (remember that’s the one with no relationships between the variables) – Not necessarily the best indices

Fit Indices Parsimony-adjusted index – These include penalties for model complexity (which normally gives you better fix by adding more paths) – These will have smaller values for simpler models

Fit Indices Predictive fix indices – Estimate model fit in a hypothetical replication of the study with the same sample size randomly drawn from the population – Not always used

Fit Indices What size? I need a rule?! – Everyone cites Hu and Bentler (1999) for the golden standards. – Same problem that Cohen had (we love rules). So when the fit is messy, cite Kline (page 197) as reasons that’s not a bad thing – This section is an interesting read, especially if you have trouble publishing, but not crucial to your understanding of fit indices

Model Test Statistic Chi-square (listed as CMIN in your output) – Formula = (N-1)F ML – F ML = is the minimum fit function in ML estimation – P values are based on df for your model and a chi- square distribution – You want this to be nonsignificant. But this is a catch 22!

Model Test Statistic Chi-square is biased by – Multivariate non-normality – Correlation size – bigger correlations can be bad for you (harder to estimate all that variance) – Unique variance – Sample size

Model Test Statistic Everyone reports chi-square, but people tend to ignore significant values – (I’m sort of eh on his YOU MUST PAY ATTN OR DIE talk in this section)

Model Test Statistic Normed chi-square or (X 2 /df) – this used to be widely reported and used – The criterion was < 3.00 were good models – Now most people have moved away from this procedure

Approximate Fit Indices Some examples: – RMSEA (root mean square error of approximation) – SRMR (standardized root mean square residual) – A/GFI (adjusted/goodness of fit index) – CFI (comparative fix index) – TLI (Tucker-Lewis Index) – NFI (Normed Fit Index)

RMSEA Parsimony-adjusted index Want small values – Excellent <.06 (not a typo different than book) – Good <.08 – Acceptable <.10 – Eeek >.10 Report CI!

Pclose Tests if the RMSEA is in the excellent range You want p >.50 to show that there is a high probability that RMSEA is effectively zero

SRMR Parsimony-adjusted index Want small values – Excellent <.06 (not a typo different than book) – Good <.08 – Acceptable <.10 – Eeek >.10

GFI Do not use this sucker unless you want to get a nasty review. – GFI, AGFI, PGFI Lots of research showing it’s positively biased Want large values

CFI Incremental Fit Index – Values are 0 to 1 (sometimes you’ll get slightly over 1, usually indicates something is wrong) – Want high values Excellent >.95 Good >.90 Blah <.90

The Other FIs NFI – a variation of the CFI, as it was said to underestimate for small samples RFI (relative fit index) IFI (incremental fit index) TLI (Tucker Lewis Index) – All have the same basic rules and formulas as CFI See Tabachnick for how these and the next slides are calculated

Some other statistics Pratio – parsimony index PNFI, PCFI are parsimony adjustments for NFI, CFI NCP – noncentrality parameter (tells you how much it leans from the normal for that distribution)

Some other statistics FMIN – minimum discrepancy function used to calculate chi-square and other statistics – Include confidence interval in Amos

Model Comparisons Let’s say you want to adjust your model – You can compare the adjusted model to the original model to determine if the adjustment is better Let’s say you want to compare two different models – You can compare their fits to see which is better

Model Comparisons Nested models – If you can create one model from another by the addition or subtraction of parameters, then it is nested Model A is said to be nested within Model B, if Model B is a more complicated version of Model A. – For example, a one-factor model is nested within a two-factor as a one-factor model can be viewed as a two-factor model in which the correlation between factors is perfect).

Nested Models Chi-square difference test – | Subtract Model 1 CMIN – Model 2 CMIN | – Subtract Model 1 df – Model 2 df – Use a chi-square table to look up p <.05 for difference in df – See if the first step is greater than that value If yes, you say the model with the lower chi-square is better If no, you say they are the same and go with the simpler model

Nested Models So how can I tell what to change? NOTE: JUST CHANGE ONE THING AT A TIME! Use modification indices! – They tell you what the chi-square change would be if you add the path suggested. – Based on X 2 (1) – called a Lagrange Multiplier Remember that p 4. You can change this to see fewer options if you have a lot.

Non-Nested Models AIC – Akaike Information Criterion – Related CAIC (consistent AIC) BIC – Bayesian Information Criterion BCC – Browne-Cudeck Criterion – All of these are how much the sample will cross validate in the future – You want them to be small, so you pick the smallest one of the two models (how different?)

Non-Nested Models ECVI – expected cross validation index MECVI – modified ECVI – Again, you want small values, so you pick the model with the smallest ECVI

OMG! So what to do? – Mainly people report: X 2, RMSEA, SRMR, CFI – Determine the type of model change to use the right model comparison statistic