Research Hypotheses and Multiple Regression: 2 Comparing model performance across populations Comparing model performance across criteria.

Slides:



Advertisements
Similar presentations
Bivariate &/vs. Multivariate
Advertisements

Transformations & Data Cleaning
Canonical Correlation simple correlation -- y 1 = x 1 multiple correlation -- y 1 = x 1 x 2 x 3 canonical correlation -- y 1 y 2 y 3 = x 1 x 2 x 3 The.
Lesson 10: Linear Regression and Correlation
Survey of Major Correlational Models/Questions Still the first & most important is picking the correct model… Simple correlation questions simple correlation.
Automated Regression Modeling Descriptive vs. Predictive Regression Models Four common automated modeling procedures Forward Modeling Backward Modeling.
Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
Chapter Thirteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved. Linear Regression and Correlation.
Research Hypotheses and Multiple Regression Kinds of multiple regression questions Ways of forming reduced models Comparing “nested” models Comparing “non-nested”
PSY 307 – Statistics for the Behavioral Sciences
Regression Models w/ k-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Interactions in Regression.
Multiple Regression Models: Some Details & Surprises Review of raw & standardized models Differences between r, b & β Bivariate & Multivariate patterns.
Multiple Regression Models
Intro to Statistics for the Behavioral Sciences PSYC 1900
Simple Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas –quantitative vs. binary predictor.
Power Analysis for Correlational Studies Remember that both “power” and “stability” are important! Useful types of power analyses –simple correlations.
Simple Correlation Scatterplots & r Interpreting r Outcomes vs. RH:
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
Simple Correlation RH: correlation questions/hypotheses and related statistical analyses –simple correlation –differences between correlations in a group.
K-group ANOVA & Pairwise Comparisons ANOVA for multiple condition designs Pairwise comparisons and RH Testing Alpha inflation & Correction LSD & HSD procedures.
An Introduction to Classification Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Relationships Among Variables
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
McGraw-Hill © 2006 The McGraw-Hill Companies, Inc. All rights reserved. Correlational Research Chapter Fifteen.
Bivariate Linear Regression. Linear Function Y = a + bX +e.
Multiple Sample Models James G. Anderson, Ph.D. Purdue University.
Power Analysis Determining Sample Size for “the study”
Understanding Statistics
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Correlation and Regression PS397 Testing and Measurement January 16, 2007 Thanh-Thanh Tieu.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Correlational Research Chapter Fifteen Bring Schraw et al.
Statistical Control Statistical Control is about –underlying causation –under-specification –measurement validity –“Correcting” the bivariate relationship.
ANOVA and Linear Regression ScWk 242 – Week 13 Slides.
Part IV Significantly Different: Using Inferential Statistics
Correlation & Regression Chapter 5 Correlation: Do you have a relationship? Between two Quantitative Variables (measured on Same Person) (1) If you have.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Discriminant Analysis Discriminant analysis is a technique for analyzing data when the criterion or dependent variable is categorical and the predictor.
Regression Models w/ 2 Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting the regression.
Advanced Meta-Analyses Heterogeneity Analyses Fixed & Random Efffects models Single-variable Fixed Effects Model – “Q” Wilson Macro for Fixed Efffects.
Chapter 16 Data Analysis: Testing for Associations.
Chapter 13 Multiple Regression
Regression Models w/ 2-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting.
Chapter 3: Statistical Significance Testing Warner (2007). Applied statistics: From bivariate through multivariate. Sage Publications, Inc.
Correlation and Regression: The Need to Knows Correlation is a statistical technique: tells you if scores on variable X are related to scores on variable.
Correlation & Regression Analysis
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Simple Correlation Scatterplots & r –scatterplots for continuous - binary relationships –H0: & RH: –Non-linearity Interpreting r Outcomes vs. RH: –Supporting.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Chapter 8 Relationships Among Variables. Outline What correlational research investigates Understanding the nature of correlation What the coefficient.
Chapter 22 Inferential Data Analysis: Part 2 PowerPoint presentation developed by: Jennifer L. Bellamy & Sarah E. Bledsoe.
1 G Lect 10M Contrasting coefficients: a review ANOVA and Regression software Interactions of categorical predictors Type I, II, and III sums of.
Methods of Presenting and Interpreting Information Class 9.
Topics: Multiple Regression Analysis (MRA)
Regression and Correlation
Understanding and conducting
Chapter 6 Predicting Future Performance
Week 3 Class Discussion.
Quantitative Methods Simple Regression.
Product moment correlation
Inferential Statistics
Psych 231: Research Methods in Psychology
Chapter 6 Predicting Future Performance
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Research Hypotheses and Multiple Regression: 2 Comparing model performance across populations Comparing model performance across criteria

Comparing model performance across groups This involves the same basic idea as comparing a bivariate correlation across groups only now we’re working with multiple predictors in a multivariate model This sort of analysis has multiple important uses … theoretical – different behavioral models for different groups? psychometric – important part of evaluating if “measures” are equivalent for different groups (such as gender, race, across cultures or within cultures over time) is unraveling the multivariate relationships among measures & behaviors applied – prediction models must not be “biased”

Comparing model performance across groups There are three different questions involved in this comparison … Does the predictor set “work better” for one group than another? Asked by comparing R 2 of predictor set from the 2 groups ? we will build a separate model for each group (allowing different regression weights for each group) then use Fisher’s Z-test to compare the resulting R 2 s Are the models “substitutable”? use a cross-validation technique to compare the models use Steiger’s t-test to compare R 2 of “direct” & “crossed” models Are the regression weights of the 2 groups “different” ? use Z-tests to compare the weights predictor-by- predictor or using interaction terms to test for group differences

Things to remember when doing these tests!!! the more collinear the variables being substituted, the more collinear they will be -- for this reason there can be strong collinearity between two models that share no predictors the weaker the two models (lower R²), the less likely they are to be differentially correlated with the criterion nonnill-H0: tests are possible -- and might be more informative!! these are not very powerful tests !!! compared to avoiding a Type II error when looking for a given r, you need nearly twice the sample size to avoid a Type II error when looking for an r-r of the same magnitude these tests are also less powerful than tests comparing nested models So, be sure to consider sample size, power and the magnitude of the r-difference between the non-nested models you compare !

Group #1 (larger n) “direct model” R² D1 y’ 1 = b 1 x + b 1 z + a 1 “direct model” R² D2 y’ 2 = b 2 x + b 2 z + a 2 Group #2 (smaller n) Comparing multiple regression models across groups  3 ?s Does the predictor set “work better” for one group than another? Compare R² D1 & R² D2 using Fisher’s Z-test Retain H0: predictor set “works equally” for 2 groups Reject H0: predictor set “works better” for higher R2 group Remember!! We are comparing the R2 “fit” of the models… But, be sure to use R in the computator!!!!

Group #1 (larger n) “G 1 direct model” R² D1 y’ 1 = b 1 x + b 1 z + a 1 “G 1 crossed model” R² X1 y’ 1 = b 2 x + b 2 z + a 2 using Hotelling’s t-test or Steiger’s Z-test will need r DX -- correlation between models – from each group Group #2 (smaller n) Are the multiple regression models “substitutable” across groups? Apply the model (bs & a) from Group 2 to the data from Group 1 “G 1 crossed model” R² X2 y’ 1 = b 2 x + b 2 z + a 2 Apply the model (bs & a) from Group 1 to the data from Group 2 “ G 2 direct model” R² D2 y’ 2 = b 2 x + b 2 z + a 2 Compare R² D2 & R² X2 Compare R² D1 & R² X1

Are the regression weights of the 2 groups “different” ? test of an interaction of predictor and grouping variable Z-tests using pooled standard error terms Asking if a single predictor has a different regression weight for two different groups is equivalent to asking if there is an interaction between that predictor and group membership. (Please note that asking about a regression slope difference and about a correlation difference are two different things – you know how to use Fisher’s Test to compare correlations across groups) This approach uses a single model, applied to the full sample… Criterion’ = b 1 predictor + b 2 group + b 3 predictor*group + a If b 3 is significant, then there is a difference between then predictor regression weights of the two groups.

However, this approach gets cumbersome when applied to models with multiple predictors. With 3 predictors we would look at the model … y’ = b 1 G + b 2 P1 + b 3 G*P1 + b 4 P2 + b 5 G*P2 + b 6 P3 + b 7 G*P3 +a Each interaction term is designed to tell us if a particular predictor has a regression slope difference across the groups. Because the collinearity among the interaction terms and between a predictor’s term and other predictor’s interaction terms all influence the interaction b weights, there has been dissatisfaction with how well this approach works for multiple predictors. Also, because his approach does not involve constructing different models for each group, it does not allow… the comparison of the “fit” of the two models an examination of the “substitutability” of the two models

Another approach is to apply a significance test to each predictor’s b weights from the two models – to directly test for a significant difference. (Again, this is different from comparing the same correlation from 2 groups). The most common formula is … b G1 - b G2 Z = SE b-difference However, there are competing formulas for “SE b-difference “ The most common formula (e.g., Cohen, 1983) is… (df bG1 * SE bG1 2 ) + (df bG2 * SE bG2 2 ) SE b-difference = √ df bG1 + df bG2

However, work by two research groups has demonstrated that, for large sample studies (both N > 30) this Standard Error estimator is negatively biased (produces error estimates that are too small), so that the resulting Z-values are too large, promoting Type I & Type 3 errors. Brame, Paternost, Mazerolle & Piquero (1998) Clogg, Petrova & Haritou (1995) Leading to the formulas … SE b-difference = √ ( SE bG1 2 + SE bG2 2 ) and… b G1 - b G2 Z = √ ( SE bG1 2 + SE bG2 2 )

Match the question with the most direct test… Practice is better correlated to performance for novices than for experts. The structure of a model involving practice, motivation & recent experience is different for novices than experts. Practice has a larger regression weight in the model for novices than for experts. Practice contributes to the regression model for novices, but not for experts. A model involving practice, motivation & recent experience better predicts performance of novices than experts. Practice is correlated with performance for novices, but not for experts. Comparing r across groups Comparing b across groups Testing r for each group Testing b for each group Comparing R 2 across groups Comparing R 2 of direct & crossed models

Comparing model performance across criteria same basic idea as comparing correlated correlations, but now the difference between the models is the criterion, not the predictor. There are two important uses of this type of comparison theoretical/applied -- do we need separate models to predict related behaviors? psychometric -- do different measures of the same construct have equivalent models (i.e., measure the same thing) ? the process is similar to testing for group differences, but what changes is the criterion that is used, rather than the group that is used we’ll apply the Hotelling’s t-test and/or Steiger’s Z-test to compare the structure of the two models

Criterion #1 “A” “A direct model” R² DA A’ = bx + bx + a “B direct model” R² DB B’ = bx + bx + a “A crossed model” R² XA A’ = bx + bx + a using Hotelling’s t-test or Steiger’s Z-tests (will need r DX -- r between models) Criterion #2 “B” Apply the model (bs & a) from group 2 to the data from group 1 Retaining the H0: for each suggests group comparability in terms of the “structure” of a single model for the two criterion variables -- there is no direct test of the differential “fit” of the two models to the two criteria. Compare R² DA & R² XA Are multiple regression models “substitutable” across criteria? Apply the model (bs & a) from group 1 to the data from group 2 “B crossed model” R² XB B’ = bx + bx + a Compare R² DB & R² XB