Outline of Today’s Discussion 1.Seeing the big picture in MR: Prediction 2.Starting SPSS on the Different Models: Stepwise versus Hierarchical 3.Interpreting.

Slides:



Advertisements
Similar presentations
Correlation and Linear Regression.
Advertisements

Prediction with multiple variables Statistics for the Social Sciences Psychology 340 Spring 2010.
Regression single and multiple. Overview Defined: A model for predicting one variable from other variable(s). Variables:IV(s) is continuous, DV is continuous.
© McGraw-Hill Higher Education. All Rights Reserved. Chapter 2F Statistical Tools in Evaluation.
MULTIPLE REGRESSION. OVERVIEW What Makes it Multiple? What Makes it Multiple? Additional Assumptions Additional Assumptions Methods of Entering Variables.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Statistics for the Social Sciences Psychology 340 Spring 2005 Prediction cont.
Statistics for the Social Sciences
Comparing the Various Types of Multiple Regression
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
© Copyright 2000, Julia Hartman 1 An Interactive Tutorial for SPSS 10.0 for Windows © by Julia Hartman Multiple Linear Regression Next.
Lecture 6: Multiple Regression
Prediction/Regression
Reminders  HW2 due today  Exam 1 next Tues (9/27) – Ch 1-5 –3 sections: Short answers (concepts, definitions) Calculations (you’ll be given the formulas)
Ch 11: Correlations (pt. 2) and Ch 12: Regression (pt.1) Nov. 13, 2014.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Lecture 5: Simple Linear Regression
Multiple Regression – Basic Relationships
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Simple Linear Regression Analysis
SPSS Statistical Package for Social Sciences Multiple Regression Department of Psychology California State University Northridge
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Split Sample Validation General criteria for split sample validation Sample problems.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Multiple Regression.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
The Goal of MLR  Types of research questions answered through MLR analysis:  How accurately can something be predicted with a set of IV’s? (ex. predicting.
1 Psych 5510/6510 Chapter 10. Interactions and Polynomial Regression: Models with Products of Continuous Predictors Spring, 2009.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Then click the box for Normal probability plot. In the box labeled Standardized Residual Plots, first click the checkbox for Histogram, Multiple Linear.
Review of Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
Multiple Linear Regression Partial Regression Coefficients.
Regression Chapter 16. Regression >Builds on Correlation >The difference is a question of prediction versus relation Regression predicts, correlation.
Fundamental Statistics in Applied Linguistics Research Spring 2010 Weekend MA Program on Applied English Dr. Da-Fu Huang.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 12, 2013 Correlation and Regression.
Regression Lesson 11. The General Linear Model n Relationship b/n predictor & outcome variables form straight line l Correlation, regression, t-tests,
Welcome to MM570 Psychological Statistics Unit 4 Seminar Dr. Srabasti Dutta.
Multiple Regression INCM 9102 Quantitative Methods.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Outline of Today’s Discussion
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Regression. Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression.
Outline of Today’s Discussion 1.Regression Analysis: Introduction 2.An Alternate Formula For Regression 3.Correlation, Regression, and Statistical Significance.
Using SPSS Note: The use of another statistical package such as Minitab is similar to using SPSS.
Welcome to MM570 Psychological Statistics Unit 4 Seminar Dr. Bob Lockwood.
Applied Quantitative Analysis and Practices LECTURE#28 By Dr. Osman Sadiq Paracha.
Week of March 23 Partial correlations Semipartial correlations
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 18 Multivariate Statistics.
Independent Samples ANOVA. Outline of Today’s Discussion 1.Independent Samples ANOVA: A Conceptual Introduction 2.The Equal Variance Assumption 3.Cumulative.
(Slides not created solely by me – the internet is a wonderful tool) SW388R7 Data Analysis & Compute rs II Slide 1.
Multiple Regression Scott Hudson January 24, 2011.
Chapter 11 REGRESSION Multiple Regression  Uses  Explanation  Prediction.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
An Interactive Tutorial for SPSS 10.0 for Windows©
Chapter 15 Multiple Regression Model Building
Statistics for the Social Sciences
Multiple Regression.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Regression.
بحث في التحليل الاحصائي SPSS بعنوان :
Statistics for the Social Sciences
Simple Linear Regression
Prediction/Regression
Prediction/Regression
Multiple Regression – Split Sample Validation
Regression Analysis.
Presentation transcript:

Outline of Today’s Discussion 1.Seeing the big picture in MR: Prediction 2.Starting SPSS on the Different Models: Stepwise versus Hierarchical 3.Interpreting the SPSS output box by box

Part 1 Seeing the Big Picture in MR: Prediction!

Seeing the Big MR Picture: Prediction No matter how hairy the output looks, just remember that MR is always about prediction! Better still, for our purposes, MR will always be about linear prediction, which is easier than some of the non-linear regression we’ve already attempted. Let’s re-visit the expanded version of our simplest linear equation:y = mx + b…

Seeing the Big MR Picture: Prediction In the world of raw scores, the equation w/b this: y = m 1 (x 1 ) + m 2 (x 2 ) + m 3 (x 3 ) …. + b Or, it may be simpler to ‘start’ with the intercept y = b + m 1 (x 1 ) + m 2 (x 2 ) + m 3 (x 3 ) ….

Seeing the Big MR Picture: Prediction In the world of raw scores, the equation w/b this: y = m 1 (x 1 ) + m 2 (x 2 ) + m 3 (x 3 ) …. + b Or, it may be simpler to ‘start’ with the intercept y = b + m 1 (x 1 ) + m 2 (x 2 ) + m 3 (x 3 ) …. Remember that this is the clinically/scientifically relevant variable we’d like to predict! Remember that SPSS calls the y-intercept “Constant”. Remember that SPSS calls the raw-score coefficients “unstandardized coefficients”.

Seeing the Big MR Picture: Prediction In the world of z-scores, the equation w/b this: Z y =  1 (Z x1 ) +  2 (Z x2 ) +  3 (Z x3 ) …. The intercept is always 0 in Z-score regression! The row in SPSS for ‘constants’ will be blank.

Seeing the Big MR Picture: Prediction In the world of z-scores, the equation w/b this: Z y =  1 (Z x1 ) +  2 (Z x2 ) +  3 (Z x3 ) …. The intercept is always 0 in Z-score regression! The row in SPSS for ‘constants’ will be blank. Remember that this is the clinically/scientifically relevant variable we’d like to predict! Remember that SPSS calls the y-intercept “Constant”. Remember that SPSS calls the z-score coefficients “standardized coefficients”.

Seeing the Big MR Picture: Prediction The coefficients for raw- or z-score prediction will come from the ‘coefficients box’ in the SPSS output. More on that later.

Part 2 Stepwise MR in SPSS: An Overview

MR Example Variables Crednow – the number of credits that a student is taking while enrolled in a stat course. Age – student’s age EmpHours – the # of hours that the student works Night – whether the student is also enrolled in night courses (yes or no) Othact – involvement in other activities (yes/no) Criterion Predictors Let’s define what the variables mean…

Stepwise MR in SPSS In SPSS, the MR process begins the same way as for simple, one-variable regression. Analyze  Regression  Linear These steps in SPSS w/b used no matter whether the procedure is stepwise or hierarchical (or other MR procedures, such as ‘simultaneous’).

Stepwise MR in SPSS Identify a single dependent variable (criterion), and more than one independent variable (predictors).

Stepwise MR in SPSS The Method box s/b set to Stepwise. Then click on the ‘Statistics’ box…

Starting SPSS on Different Models Click on the ‘Statistics’ box to choose your statistics FOR ANY MR MODEL

Stepwise MR in SPSS We probably won’t need all the statistics options, but let’s select them all for now.

Stepwise MR in SPSS When your variables, method, and stats are complete, click OK to run the multiple regression!

Interpreting The Output: Box by Box These are the ‘zero-order’ correlations. These do NOT depend on which MR procedure is used. The Correlations Box Could Be Helpful! Stepwise MR in SPSS

Interpreting The Output: Box by Box IMO, The Variables Entered/Removed Box Is NOT particularly Helpful! Notice that there are 3 different models being constructed. Successive models have more variables entered. Or…more variables removed if you start with the full model and exclude variables.

Interpreting The Output: Box by Box The Model Summary Box Is EXTREMELY Helpful! Notice that there are 3 different models being constructed. You can see the INCREMENTAL value of each model by comparing r-squared values, or r-square-change values! The ‘footnotes’ tell you everything that the ‘variables entered/removed’ box does.

Interpreting The Output: Box by Box IMO, The ANOVA Box Is NOT particularly Helpful! This box would only be helpful if you needed the ‘sum of square’ or ‘mean square’ values. That’s unlikely!

Interpreting The Output: Box by Box The Coefficients Box Is the MOST Helpful for Prediction! Notice that there are 3 different models being constructed. The box provides coefficients for each model! And, it does this for raw scores (unstandardized ‘B’), and z-scores (standardized Beta)

Interpreting The Output: Box by Box IMO, The Excluded Variables Box Is NOT particularly Helpful! The information here is largely (not entirely) redundant with that of the other boxes.

Building Prediction Equations: Coefficients are Our Friends!

Building the Prediction Equations We can use the coefficients box to build our regression equations. Remember, we can do regression equations using z-scores OR raw scores. The z-score equations make more sense intuitively b/c the standardized beta values (standardized slopes) allow for direct comparisons across predictors that use different units of measure.

Building the Prediction Equations Stepwise MR We can build an equation for any of the three models shown here. Model #3 will be the most complete. Raw-score model first… The coefficients box helps us to build the regression equations.

Building the Prediction Equations Stepwise MR Raw Score: CredNow = – 3.42(night) -.087(age) (emphours) The coefficients box helps us to build the regression equations.

Building the Prediction Equations Stepwise MR Now, the z-score model… Z-Score: Z CredNow = 0 –.374(Z night ) -.205(Z age ) -.186(Z emphours ) The coefficients box helps us to build the regression equations.

Building the Prediction Equations Stepwise MR Let’s go back to the raw-score equation and predict a CREDNOW score…. Assume that a particular student’s NIGHT score = 1 (s/he is a night student) And their AGE = 50 years old And their EMPHOURS = 40 hours per week The coefficients box helps us to build the regression equations.

Building the Prediction Equations Stepwise MR Raw Score: CredNow = – 3.42(1) -.087(50) (40) Raw Score: CredNow = – – Predicted Raw Score = credits now The coefficients box helps us to build the regression equations.

Part 2 Hierarchical MR in SPSS: An Overview

Hierarchical MR in SPSS In SPSS, the MR process begins the same way as for simple, one-variable regression. Analyze  Regression  Linear These steps in SPSS w/b used no matter whether the procedure is stepwise or hierarchical (or other MR procedures, such as ‘simultaneous’).

Hierarchical MR in SPSS Identify a single dependent variable (criterion), and more than one independent variable (predictors).

Hierarchical MR in SPSS The Method box s/b set to ENTER. Note: There is no ‘hierarchical’ option, per se. Then click on the ‘Statistics’ box…

Starting SPSS on Different Models Within the Method box, there is no option called “hierarchical”. We have to “ENTER” our hierarchy of variables manually, using the previous and next buttons to specify the order of blocks. Example: Block 1 of 1 Example: Block 2 of 2 Example: Block 3 of 3

Hierarchical MR in SPSS We probably won’t need all the statistics options, but let’s select them all for now.

Hierarchical MR in SPSS When your variables, method, and stats are complete, click OK to run the multiple regression!

Hierarchical MR in SPSS These are the ‘zero-order’ correlations. These do NOT depend on which MR procedure is used.

Interpreting The Output: Box by Box So, here are the three most helpful boxes (“CMC”) Answers the question: What’s correlated with what? Answers the question: How good are the models? Answers the question: What are my coefficients? (Used for prediction… the big picture!) Correlations Model Summary Coefficients

Part 3 Interpreting Differences Between the Two Models

Differences Between The Models We can compare the two models in several ways. First, we might consider the models within the context of “Ockham’s Razor”… In other words, how much variance does each model explain ‘per unit complexity’? That is, fewer variables w/b better ALL OTHER THINGS BEING EQUAL.

Differences Between The Models Hierarchical MR Model Stepwise MR Model

Differences Between The Models We can also compare the regression equations for each type of model. Remember, we can do regression equations using z-scores OR raw scores. The z-score equations make more sense intuitively b/c the standardized beta values (standardized slopes) allow for direct comparisons across predictors that use different units of measure.

Differences Between The Models Stepwise MR Raw Score: CredNow = – 3.42(night) -.087(age) (emphours) Z-Score: Z CredNow = 0 –.374(Z night ) -.205(Z age ) -.186(Z emphours )

Differences Between The Models Hierarchical MR Raw Score: CredNow = – 3.31(night) (age) (emphours) -.806(othact) Z-Score: Z CredNow = 0 –.362(Z night ) -.183(Z age ) -.186(Z emphours ) -.111(Z othact )

Differences Between The Models Stepwise (r 2 = 0.355) with 3 variables Raw Score: CredNow = – 3.42(night) -.087(age) (emphours) Hierarchical(r 2 = 0.367) with 4 variables Raw Score: CredNow = – 3.31(night) (age) (emphours) -.806(othact) Here are the ‘competing’ Raw Score Regression Models

Differences Between The Models Stepwise (r 2 = 0.355) with 3 variables Z-Score: ZCredNow = 0 –.374(Z night ) -.205(Z age ) -.186(Z emphours ) Hierarchical(r 2 = 0.367) with 4 variables Z-Score: Z CredNow = 0 –.362(Z night ) -.183(Z age ) -.186(Z emphours ) -.111(Z othact ) Here are the ‘competing’ Z-Score Regression Models