Regression Models w/ 2-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting.

Slides:



Advertisements
Similar presentations
Canonical Correlation simple correlation -- y 1 = x 1 multiple correlation -- y 1 = x 1 x 2 x 3 canonical correlation -- y 1 y 2 y 3 = x 1 x 2 x 3 The.
Advertisements

Getting More out of Multiple Regression Darren Campbell, PhD.
Automated Regression Modeling Descriptive vs. Predictive Regression Models Four common automated modeling procedures Forward Modeling Backward Modeling.
Research Hypotheses and Multiple Regression: 2 Comparing model performance across populations Comparing model performance across criteria.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
Analytic Comparisons & Trend Analyses Analytic Comparisons –Simple comparisons –Complex comparisons –Trend Analyses Errors & Confusions when interpreting.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Linear Discriminant Function LDF & MANOVA LDF & Multiple Regression Geometric example of LDF & multivariate power Evaluating & reporting LDF results 3.
Regression Models w/ k-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting.
Multiple Group X² Designs & Follow-up Analyses X² for multiple condition designs Pairwise comparisons & RH Testing Alpha inflation Effect sizes for k-group.
Introduction to the General Linear Model (GLM) l 1 quantitative variable & 1 2-group variable l 1a  main effects model with no interaction l 1b  interaction.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
Plots, Correlations, and Regression Getting a feel for the data using plots, then analyzing the data with correlations and linear regression.
Multiple Regression Models: Some Details & Surprises Review of raw & standardized models Differences between r, b & β Bivariate & Multivariate patterns.
Analyses of K-Group Designs : Analytic Comparisons & Trend Analyses Analytic Comparisons –Simple comparisons –Complex comparisons –Trend Analyses Errors.
Introduction the General Linear Model (GLM) l what “model,” “linear” & “general” mean l bivariate, univariate & multivariate GLModels l kinds of variables.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Linear Regression and Correlation Analysis
MANOVA LDF & MANOVA Geometric example of MANVOA & multivariate power MANOVA dimensionality Follow-up analyses if k > 2 Factorial MANOVA.
Multivariate Analyses & Programmatic Research Re-introduction to Multivariate research Re-introduction to Programmatic research Factorial designs  “It.
Lecture 27 Polynomial Terms for Curvature Categorical Variables.
General Linear Models -- #1 things to remember b weight interpretations 1 quantitative predictor 1 quantitative predictor & non-linear component 1 2-group.
Simple Correlation Scatterplots & r Interpreting r Outcomes vs. RH:
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
Elaboration Elaboration extends our knowledge about an association to see if it continues or changes under different situations, that is, when you introduce.
Simple Correlation RH: correlation questions/hypotheses and related statistical analyses –simple correlation –differences between correlations in a group.
Ch. 14: The Multiple Regression Model building
Introduction the General Linear Model (GLM) l 2 quantitative variables l 3a  main effects model with no interaction l 3b  interaction model l 2 categorical.
2x2 BG Factorial Designs Definition and advantage of factorial research designs 5 terms necessary to understand factorial designs 5 patterns of factorial.
An Introduction to Classification Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification.
CORRELATIONAL RESEARCH I Lawrence R. Gordon Psychology Research Methods I.
Correlation and Regression Analysis
Experimental Group Designs
CORRELATIO NAL RESEARCH METHOD. The researcher wanted to determine if there is a significant relationship between the nursing personnel characteristics.
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Statistical hypothesis testing – Inferential statistics II. Testing for associations.
WG ANOVA Analysis of Variance for Within- groups or Repeated measures Between Groups and Within-Groups ANOVA WG ANOVA & WG t-tests When to use each Similarities.
Moderation & Mediation
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Correlation and Regression SCATTER DIAGRAM The simplest method to assess relationship between two quantitative variables is to draw a scatter diagram.
Correlational Research Chapter Fifteen Bring Schraw et al.
Coding Multiple Category Variables for Inclusion in Multiple Regression More kinds of predictors for our multiple regression models Some review of interpreting.
Lab 5 instruction.  a collection of statistical methods to compare several groups according to their means on a quantitative response variable  Two-Way.
Multiple regression models Experimental design and data analysis for biologists (Quinn & Keough, 2002) Environmental sampling and analysis.
Regression Models w/ 2 Categorical Variables Sources of data for this model Variations of this model Definition and advantage of factorial research designs.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
Multiple Regression BPS chapter 28 © 2006 W.H. Freeman and Company.
Regression Models w/ 2 Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting the regression.
Advanced Meta-Analyses Heterogeneity Analyses Fixed & Random Efffects models Single-variable Fixed Effects Model – “Q” Wilson Macro for Fixed Efffects.
Chapter 13 Multiple Regression
Plotting Linear Main Effects Models Interpreting 1 st order terms w/ & w/o interactions Coding & centering… gotta? oughta? Plotting single-predictor models.
Analysis of Covariance Combines linear regression and ANOVA Can be used to compare g treatments, after controlling for quantitative factor believed to.
ANCOVA. What is Analysis of Covariance? When you think of Ancova, you should think of sequential regression, because really that’s all it is Covariate(s)
28. Multiple regression The Practice of Statistics in the Life Sciences Second Edition.
Meta-Analysis Effect Sizes effect sizes – r, d & OR computing effect sizes estimating effect sizes & other things of which to be careful!
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, partial correlations & multiple regression Using model plotting to think about ANCOVA & Statistical control Homogeneity.
ANCOVA.
General Linear Model What is the General Linear Model?? A bit of history & our goals Kinds of variables & effects It’s all about the model (not the data)!!
Week of March 23 Partial correlations Semipartial correlations
Biostatistics Regression and Correlation Methods Class #10 April 4, 2000.
Linear Discriminant Function Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification & LDF.
29 October 2009 MRC CBU Graduate Statistics Lectures 4: GLM: The General Linear Model - ANOVA & ANCOVA1 MRC Cognition and Brain Sciences Unit Graduate.
Appendix I A Refresher on some Statistical Terms and Tests.
1. Refresher on the general linear model, interactions, and contrasts UCL Linguistics workshop on mixed-effects modelling in R May 2016.
Stats Methods at IC Lecture 3: Regression.
Plotting Non-linear & Complex Main Effects Models
Introduction to the General Linear Model (GLM)
Quantitative Methods Simple Regression.
Soc 3306a Lecture 11: Multivariate 4
Presentation transcript:

Regression Models w/ 2-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting the regression weight –Plotting and interpreting the model Interaction version of the model –Composing the interaction term –Testing the interaction term = testing homogeneity of regression slope assumption –Interpreting the regression weight –Plotting and interpreting the model Plotting more complex models

As always, “the model doesn’t care where the data come from”. Those data might be … a measured binary variable (e.g., ever- vs. never-married) and a measured quant variable (e.g., age) a manipulated binary variable (Tx vs. Cx) and a measured quant variable (e.g., age) a measured binary variable (e.g., ever- vs. never-married) and a manipulated quant variable (e.g., 0, 1, 2, 5, or 10 practices) a manipulated binary variable (Tx vs. Cx) and a manipulated quant variable (e.g., 0, 1, 2, 5, or 10 practices) Like nearly every model in the ANOVA/regression/GLM family – this model was developed for and originally applied to experimental designs with the intent of causal interpretability !!! As always, causal interpretability is a function of design (i.e., assignment, manipulation & control procedures) – not statistical model or the constructs involved !!!

There are two important variations of this model 1.Main effects model Terms for the binary variable & quant variable No interaction – assumes regression slope homogeneity b-weights for binary & quant variables each represent main effect of that variable 2. Interaction model Terms for binary variable & quant variable Term for interaction - does not assume reg slp homogen !! b-weights for binary & quant variables each represent the simple effect of that variable when the other variable = 0 b-weight for the interaction term represented how the simple effect of one variable changes with changes in the value of the other variable (e.g., the extent and direction of the interaction)

a  regression constant expected value of Y if all predictors = 0 mean of the control group (G3) height of control group Y-X regression line b 2  regression weight for dummy coded binary predictor expected direction and extent of change in Y for a 1-unit increase in Z, after controlling for the other variable(s) in the model main effect of Z group Y-X regression line height difference Models with a centered quantitative predictor & a dummy coded binary predictor y’ = b 1 X + b 2 Z + a This is called a main effects model  there are no interaction terms. b 1  regression weight for centered quant predictor expected direction and extent of change in Y for a 1-unit increase in X after controlling for the other variable(s) in the model main effect of X Slope of Y-X regression line for both groups

Model  y’ = b 1 X + b 2 Z + a y’ = b 1 X + b 2 *0 + a y’ = b 1 X + a y’ = b 1 X + b 2 *1 + a y’ = b 1 X + ( b 2 + a) To plot the model we need to get separate regression formulas for each Z group. We start with the multiple regression model… For the Comparison Group coded Z = 0 Substitute the 0 in for Z Simplify the formula For the Target Group coded Z = 1 Substitute the 1 in for Z Simplify the formula slope height slope height

a b1b1 b2b2 Cx Tx  X cen a = ht of Cx line  mean of Cx b 1 = slp of Cx line b 2 = htdif Cx & Tx  Cx & Tx mean dif Cx slp = Tx slp No interaction X cen = X – X mean Z = Tx1 vs. Cx(0) y’ = b 1 X + b 2 Z + a Plotting & Interpreting Models with a centered quantitative predictor & a dummy coded binary predictor This is called a main effects model  no interaction  the regression lines are parallel.

a -b 1 b 2 = 0 Cx Tx  X cen X cen = X – X mean Z = Tx1 vs. Cx(0) a = ht of Cx line  mean of Cx b 1 = slp of Cx line b 2 = htdif Cx & Tx  Cx & Tx mean dif Cx slp = Tx slp No interaction y’ = -b 1 X + -b 2 Z + a Plotting & Interpreting Models with a centered quantitative predictor & a dummy coded binary predictor This is called a main effects model  no interaction  the regression lines are parallel.

a b 1 = 0 a b2b2 Cx Tx  X cen a = ht of Cx line  mean of Cx b 1 = slp of Cx line b 2 = htdif Cx & Tx  Cx & Tx mean dif Cx slp = Tx slp No interaction X cen = X – X mean Z = Tx1 vs. Cx(0) Plotting & Interpreting Models with a centered quantitative predictor & a dummy coded binary predictor y’ = b 1 X + b 2 Z + a This is called a main effects model  no interaction  the regression lines are parallel.

In SPSS you have to compute the interaction term – as the product of the binary variable dummy code & the centered quantitative variable So, if you have sex_dc (0=male & 1=female) and age_cen centered at its mean, you would compute the interaction as… compute age_sex_int = sex_dc * age_cen. males will have age_sex_int values of 0 females will have age_sex_int values = their age_cen values Models with Interactions As in Factorial ANOVA, an interaction term in multiple regression is a “non-additive combination” there are two kinds of combinations – additive & multiplicative main effects are “additive combinations” an interaction is a “multiplicative combination”

Testing the interaction/regression homogeneity assumption… There are two equivalent ways of testing the significance of the interaction term: 1.The t-test of the interaction term will tell whether or not b=0 2.A nested model comparison, using the R 2 Δ F-test to compare the main effect model (dummy-coded binary variable & centered quant variable) with the full model (also including the interaction product term) These are equivalent because t 2 = F, both with the same df & p. Retaining H0: means that the interaction term does not contribute to the model, after controlling for the main effects which can also be called regression homogeneity.

Interpreting the interaction regression weight If the interaction contributes to the model, we need to know how to interpret the regression weight for the interaction term. We are used to regression weight interpretations that read like, “The direction and extent of the expected change in Y for a 1-unit increase in X, holding all the other variables in the model constant at 0.” Remember that an interaction in a regression model is about how the slope between the criterion and one predictor is different for different values of another predictor. So, the interaction regression weight interpretation changes just a bit… An interaction regression weight tells the direction and extent of change in the slope of the Y-X regression line for each 1-unit increase in Z, holding all the other variables in the model constant at 0. Notice that in interaction is about regression slope differences, not correlation differences – you already know how to compare corrs

Interpreting the interaction regression weight, cont. Like interactions in ANOVA, interactions in multiple regression tell how the relationship between the criterion and one variable changes for different values of the other variable – i.e., how the simple effects differ. Just as with ANOVA, we can pick either variable as the simple effect, and see how the simple effect of that variable is different for different values of the other variable. The difference is that in this model, one variable is a quantitative variable (X) and the other is a binary grouping variable (Z) So, we can describe the interaction in 2 different ways – both from the same interaction regression weight! how does the Y-X regression line slope differ for the 2 groups? how does the Y-X regression line height difference differ for different values of X (how does the mean difference differ for different values of X)?

Interpreting the interaction regression weight, cont. Example: FB = feedback 0 = no feedback 1 = feedback perf’ = 8.2*#pract + 4.5*FB + 4.0Pr_FB We can describe the interaction regression weight 2 ways: 1.The expected direction and extent of change in the Y-X regression slope for each 1-unit increase in Z, holding… The slope of the performance-practice regression line for those with feedback (coded 1) has a slope 4 more than the slope of the regression line for those without feedback (coded 0). 2. The expected direction and extent of change in group mean difference for each 1-unit increase in X, holding … The mean performance difference between the feedback and no feedback groups will increase by 4 with each additional practice.

Interpreting the interaction regression weight, cont. perf’ = 8.2*#pract + 4.5*FB + 4.0Pr_FB The slope of the performance-practice regression line for those with feedback (coded 1) has a slope 4 more than the slope of the regression line for those without feedback (coded 0). Be sure to notice that it says “more” -- it doesn’t say whether both are positive, negative or one of each !!! Both of the plots below show FB with a “more positive” slope that nFB FB nFB FB nFB

a  regression constant tge expected value of Y if all predictors = 0 mean of the control group (G3) height of control group Y-X regression line b 2  regression weight for dummy coded binary predictor expected direction and extent of change in Y for a 1-unit increase in X, after controlling for the other variable(s) in the model simple effect of Z when X = 0 Y-X reg line height difference of groups when X = 0 (the centered mean) Models with a centered quantitative predictor, a dummy coded binary predictor & their interaction y’ = b 1 X + b 2 Z + b 3 XZ + a b 1  regression weight for centered quant predictor expected direction and extent of change in Y for a 1-unit increase in X, after controlling for the other variable(s) in the model simple effect of X when Z = 0 (comparison group) slope of Y-X regression line for the comparison group (Z coded 0) b 3  regression weight for interaction term expected direction and extent of change in the Y-X regression slope for each 1-unit increase in Z, after controlling for the other variable(s) in the model expected direction and extent of change in group mean difference for each 1-unit increase in X, after controlling for the other variable(s) in the model Y-X reg line slope difference of groups

Model  y’ = (b 1 + b 3 *0)X + (b 2 *0 + a) y’ = b 1 X + a To plot the model we need to get separate regression formulas for each Z group. We start with the multiple regression model… For the Comparison Group coded Z = 0 Substitute the 0 in for Z Simplify the formula For the Target Group coded Z = 1 slope height y’ = b 1 X + b 2 Z + b 3 XZ + a y’ = (b 1 + b 3 *1)X + (b 2 *1 + a) y’ = (b 1 + b 3 )X + (b 2 + a) slope height Substitute the 1 in for Z Simplify the formula y’ = b 1 X + b 3 XZ + b 2 Z + a y’ = (b 1 + b 3 Z)X + (b 2 Z + a) Gather all “Xs” together Factor out “X” slope height

a b1b1 b2b2 Cx Tx  X cen XZ = X cen * Z b3b3 X cen = X – X mean Z = Tx1 vs. Cx(0) y’ = b 1 X + b 2 Z + b 3 XZ + a a = ht of Cx line  mean of Cx b 1 = slp of Cx line b 2 = htdif Cx & Tx  Cx & Tx mean dif at X=0 b 3 = slpdif Cx & Tx Plotting Models with a centered quantitative predictor, a dummy coded binary predictor & their interaction

a b1 b1 b2b2 Cx Tx  X cen XZ = X cen * Z X cen = X – X mean Z = Tx1 vs. Cx(0) b3b3 y’ = b 1 X + b 2 Z + b 3 XZ + a a = ht of Cx line  mean of Cx b 1 = slp of Cx line b 2 = htdif Cx & Tx  Cx & Tx mean dif at X=0 b 3 = slpdif Cx & Tx Plotting Models with a centered quantitative predictor, a dummy coded binary predictor & their interaction

a b1 b1 b2b2 Cx Tx  X cen b 3 = 0 XZ = X cen * Z X cen = X – X mean Z = Tx1 vs. Cx(0) y’ = b 1 X + b 2 Z + b 3 XZ + a a = ht of Cx line  mean of Cx b 1 = slp of Cx line b 2 = htdif Cx & Tx  Cx & Tx mean dif at X=0 b 3 = slpdif Cx & Tx Plotting Models with a centered quantitative predictor, a dummy coded binary predictor & their interaction

So, what do the significance tests from this model tell us and what do they not tell us about the model we have plotted? We know whether or not the slope of the group coded 0 is = 0 (t-test of the quant variable weight). We know whether or not the slope of the group coded 1 is different from the slope of the group coded 0 (t-test of the interaction term weight) But, there is no t-test to tell us if the slope of the Y-X regression line for the group coded 1 = 0. We know whether or not the mean of the group coded 1 is different from the mean of the group coded 0, when X =0 (its mean; t-test of the binary variable weight. But, there is no test of the group mean difference at any other value of X. This is important when there is an interaction, because the interaction tells us the group means differ for different values of X.