Cause (Part II) - Causal Systems I. The Logic of Multiple Relationships II. Multiple Correlation Topics: III. Multiple Regression IV. Path Analysis.

Slides:



Advertisements
Similar presentations
Basic Statistics Correlation.
Advertisements

Solving Systems of Equations using Substitution
Example 12.2 Multicollinearity | 12.3 | 12.3a | 12.1a | 12.4 | 12.4a | 12.1b | 12.5 | 12.4b a12.1a a12.1b b The Problem.
Multiple Regression and Model Building
Canonical Correlation
Partial and Semipartial Correlation
Structural Equation Modeling
Data Analysis: Relationships Continued Regression
Statistics and Quantitative Analysis U4320
Bivariate and Partial Correlations. X (HT) Y (WT) The Graphical View– A Scatter Diagram X’ Y’
Soc 3306a: Path Analysis Using Multiple Regression and Path Analysis to Model Causality.
Regression single and multiple. Overview Defined: A model for predicting one variable from other variable(s). Variables:IV(s) is continuous, DV is continuous.
Understanding the General Linear Model
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Interactions in Regression.
Intro to Statistics for the Behavioral Sciences PSYC 1900
CHAPTER 30 Structural Equation Modeling From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach,
Elaboration Elaboration extends our knowledge about an association to see if it continues or changes under different situations, that is, when you introduce.
Crash Course in Correlation and Regression MEASURING ASSOCIATION Establishing a degree of association between two or more variables gets at the central.
Multiple Regression 2 Sociology 5811 Lecture 23 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Discriminant Analysis Testing latent variables as predictors of groups.
Relationships Among Variables
Structural Equation Models – Path Analysis
Example of Simple and Multiple Regression
Lecture 3-2 Summarizing Relationships among variables ©
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
LEARNING PROGRAMME Hypothesis testing Intermediate Training in Quantitative Analysis Bangkok November 2007.
Soc 3306a Lecture 8: Multivariate 1 Using Multiple Regression and Path Analysis to Model Causality.
Soc 3306a Lecture 10: Multivariate 3 Types of Relationships in Multiple Regression.
Solving Systems of Linear Equations
Relationships between Variables. Two variables are related if they move together in some way Relationship between two variables can be strong, weak or.
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Chapter 12 Examining Relationships in Quantitative Research Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Soc 3306a Multiple Regression Testing a Model and Interpreting Coefficients.
Statistics and Quantitative Analysis U4320 Segment 12: Extension of Multiple Regression Analysis Prof. Sharyn O’Halloran.
Soc 3306a Lecture 9: Multivariate 2 More on Multiple Regression: Building a Model and Interpreting Coefficients.
Chapter 17 Partial Correlation and Multiple Regression and Correlation.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
BIOL 582 Lecture Set 11 Bivariate Data Correlation Regression.
Path Analysis. Remember What Multiple Regression Tells Us How each individual IV is related to the DV The total amount of variance explained in a DV Multiple.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Statistical analysis Outline that error bars are a graphical representation of the variability of data. The knowledge that any individual measurement.
September 18-19, 2006 – Denver, Colorado Sponsored by the U.S. Department of Housing and Urban Development Conducting and interpreting multivariate analyses.
Examining Relationships in Quantitative Research
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
1 B IVARIATE AND MULTIPLE REGRESSION Estratto dal Cap. 8 di: “Statistics for Marketing and Consumer Research”, M. Mazzocchi, ed. SAGE, LEZIONI IN.
Correlation & Regression Analysis
Residuals Recall that the vertical distances from the points to the least-squares regression line are as small as possible.  Because those vertical distances.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Regression. Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression.
DTC Quantitative Research Methods Regression I: (Correlation and) Linear Regression Thursday 27 th November 2014.
Using SPSS Note: The use of another statistical package such as Minitab is similar to using SPSS.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
Biostatistics Regression and Correlation Methods Class #10 April 4, 2000.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Multiple Independent Variables POLS 300 Butz. Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality.
Statistical analysis.
Regression Analysis.
Statistical analysis.
Multiple Regression.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Cause (Part II) - Causal Systems
Product moment correlation
Quote of the Day "Not everything that counts can be counted and not everything that can be counted counts". Albert Einstein.
Regression Analysis.
Chapter 14 Multiple Regression
Presentation transcript:

Cause (Part II) - Causal Systems I. The Logic of Multiple Relationships II. Multiple Correlation Topics: III. Multiple Regression IV. Path Analysis

Cause (Part II) - Causal Systems Y X2 X1 One Dependent Variable, Multiple Independent Variables In this diagram the overlap of any two circles can be thought of as the r 2 between the two variables. When we add a third variable, however, we must ‘partial out’ the redundant overlap of the additional independent variables. R NR I. The Logic of Multiple Relationships

Cause (Part II) - Causal Systems II. Multiple Correlation Y X2 X1 R NR R 2 y.x 1 x 2 = r 2 yx 1 + r 2 yx 2 YX2X1 NR R 2 y.x 1 x 2 = r 2 yx 1 + r 2 yx 2.x 1 Notice that when the Independent Variables are independent of each other, the multiple correlation coefficient (R 2 ) is simply the sum of the individual r 2, but if the independent variables are related, R 2 is the sum of one zero order r 2 of one plus the partial r 2 of the other(s). This is required to compensate for the fact that multiple independent variables being related to each other would be otherwise double counted in explaining the same portion of the dependent variable. Partially out this redundancy solves this problem.

Cause (Part II) - Causal Systems II. Multiple Regression YX2X1 X2 Y Y’ = a + b yx 1 X 1 + b yx 2 X 2 Y’ = B yx 1 X 1 + B yx 2 X 2 or Standardized If we were to translate this into the language of regression, multiple independent variables, that are themselves independent of each other would have their own regression slopes and would simply appear as an another term added in the regression equation.

Cause (Part II) - Causal Systems Multiple Regression Y X2 X1 X2 Y Y’ = a + b yx 1 X 1 + b yx 2.x 1 X 2 or Standardized Y’ = B yx 1 X 1 + B yx 2.x 1 X 2 Once we assume the Independent Variables are themselves related with respect to the variance explained in the Dependent Variable, then we must distinguish between direct and indirect predictive effects. We do this using partial regression coefficients to find these direct effects. When standardized these B-values are called “Path coefficients” or “Beta Weights”

III. Path Analysis – The Steps and an Example 2. Calculate the Correlation Matrix 3. Specify the Path Diagram 4. Enumerate the Equations 1. Input the data 5. Solve for the Path Coefficients (Betas) 6. Interpret the Findings Cause (Part II) - Causal Systems

Path Analysis – Steps and Example Step1 – Input the data Y = DV - income X3 = IV - educ X2 = IV - pedu X1 = IV - pinc Assume you have information from ten respondents as to their income, education, parent’s education and parent’s income. We would input these ten cases and four variables into SPSS in the usual way, as here on the right. In this analysis we will be trying to explain respondent’s income (Y), using the three other independent variables (X1, X2, X3)

Step 2 – Calculate the Correlation Matrix X1 X2 X3 Y Path Analysis – Steps and Example These correlations are calculated in the usual manner through the “analyze”, “correlate”, bivariate menu clicks. Notice the zero order correlations of each IV with the DV. Clearly these IV’s must interrelate as the values of the r 2 would sum to an R 2 indicating more than 100% of the variance in the DV which, of course, is impossible.

Step 3 – Specify the Path Diagram Y X3 X1 X2 b c X3 = Offspring’s education X2 = Parent’s education X1 = Parent’s income Y = Offspring’s income Time a d e f Path Analysis – Steps and Example Therefore, we must specify a model that explains the relationship among the variables across time We start with the dependent variable on the right most side of the diagram and form the independent variable relationship to the left, indicating their effect on subsequent variables.

Step 4 – Enumerate the Path Equations 1. r yx1 = a + br x3x1 + cr x2x1 2. r yx2 = c + br x3x2 + ar x1x2 3. r yx3 = b + ar x1x3 + cr x2x3 4. r x3x2 = d + er x1x2 6. r x1x2 = f 5. r x3x1 = e + dr x1x2 b c a d e f X3 X1 X2 Y Path Analysis – Steps and Example Click here for solution to two equations in two unknownshere With the diagram specified, we need to articulate the formulae necessary to find the path coefficients (arbitrarily indicated here by letters on each path). Overall correlations between an independent and the dependent variable can be separated into its direct effect plus the sum of its indirect effects.

Step 5 – Solve for the Path Coefficients – a, b and c Path Analysis – Steps and Example The easiest way to calculate B is to use the Regression module in SPSS. By indicating income as the dependent variable and pinc, pedu and educ as the independent variables, we can solve for the Beta Weights or Path Coefficients for each of the Independent Variables. These circled numbers correspond to Beta for paths a, c and b, respectively, in the previous path diagram.

Step 5 – Solve for the Path Coefficients – d and e Path Analysis – Steps and Example The easiest way to calculate B is to use the Regression module in SPSS. By indicating offspring education as the dependent variable and Parents Inc and Parents Edu as the independent variables, we can solve for the Beta Weights or Path Coefficients for each of these Independent Variables on the DV Offspring Edu. These circled numbers correspond to Beta for paths d and e, respectively, in the previous path diagram.

The SPSS Regression module also calculate R 2. According to this statistic, for our data, 50% of the variation in the respondent’s income (Y) is accounted for by the respondent’s education (X3), parent’s education (X2) and parent’s income (X1) Path Analysis – Steps and Example Step 5a – Solving for R 2 R 2 is calculated by multiplying the Path Coefficient (Beta) by its respective zero order correlation and summed across all of the independent variables (see spreadsheet at right).

Checking the Findings Y X3 X1 X2 r =.57 B = = (.82) +.63(.68).52 = (.75) +.31(.82).69 = (.75) +.31(.68) Time r =.69 B =.63 r =.82 B =.58 r = B =.68 e =.50 r =.52 B = -.21 r =.75 B =.35 The values of r and B tells us three things: 1) the value of Beta is the direct effect; 2) dividing Beta by r gives the proportion of direct effect; and 3) the product of Beta and r summed across each of the variables with direct arrows into the dependent variable is R 2. The value of 1-R 2 is e. Path Analysis – Steps and Example r yx1 = a + br x3x1 + cr x2x1 r yx2 = c + br x3x2 + ar x1x2 r yx3 = b + ar x1x3 + cr x2x3

Step 6 – Interpret the Findings Y X3 X1 X X3 = Offspring’s education X2 = Parent’s education X1 = Parent’s income Y = Offspring’s income Time e =.50 Specifying the Path Coefficients (Betas), several facts are apparent, among which are that Parent’s income has the highest percentage of direct effect (i.e.,.63/.69 = 92% of its correlation is a direct effect, 8% is an indirect effect). Moreover, although the overall correlation of educ with income is positive, the direct effect of offspring’s education, in these data, is actually negative! Path Analysis – Steps and Example End

Exercise - Solving Two Equations in Two Unknowns Back