General Linear Model What is the General Linear Model?? A bit of history & our goals Kinds of variables & effects It’s all about the model (not the data)!!

Slides:



Advertisements
Similar presentations
Non-linear Components for Multiple Regression Why we might need non-linear components Type of non-linear components Squared-term components & their disadvantages.
Advertisements

Transformations & Data Cleaning
Canonical Correlation simple correlation -- y 1 = x 1 multiple correlation -- y 1 = x 1 x 2 x 3 canonical correlation -- y 1 y 2 y 3 = x 1 x 2 x 3 The.
Cluster Analysis Purpose and process of clustering Profile analysis Selection of variables and sample Determining the # of clusters.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
Factorial ANOVA More than one categorical explanatory variable.
Analytic Comparisons & Trend Analyses Analytic Comparisons –Simple comparisons –Complex comparisons –Trend Analyses Errors & Confusions when interpreting.
Intro to Factorial ANOVA
ANOVA: PART II. Last week  Introduced to a new test:  One-Way ANOVA  ANOVA’s are used to minimize family-wise error:  If the ANOVA is statistically.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Linear Discriminant Function LDF & MANOVA LDF & Multiple Regression Geometric example of LDF & multivariate power Evaluating & reporting LDF results 3.
Statistical Control Regression & research designs Statistical Control is about –underlying causation –under-specification –measurement validity –“Correcting”
Regression Models w/ k-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
January 6, afternoon session 1 Statistics Micro Mini Multiple Regression January 5-9, 2008 Beth Ayers.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
Power Analysis for Correlation & Multiple Regression Sample Size & multiple regression Subject-to-variable ratios Stability of correlation values Useful.
Multiple Regression Models: Some Details & Surprises Review of raw & standardized models Differences between r, b & β Bivariate & Multivariate patterns.
ANCOVA Psy 420 Andrew Ainsworth. What is ANCOVA?
Analyses of K-Group Designs : Analytic Comparisons & Trend Analyses Analytic Comparisons –Simple comparisons –Complex comparisons –Trend Analyses Errors.
Introduction the General Linear Model (GLM) l what “model,” “linear” & “general” mean l bivariate, univariate & multivariate GLModels l kinds of variables.
Simple Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas –quantitative vs. binary predictor.
Introduction to Path Analysis & Mediation Analyses Review of Multivariate research & an Additional model Structure of Regression Models & Path Models Direct.
Sampling & Experimental Control Psych 231: Research Methods in Psychology.
MANOVA LDF & MANOVA Geometric example of MANVOA & multivariate power MANOVA dimensionality Follow-up analyses if k > 2 Factorial MANOVA.
Introduction to Multivariate Research & Factorial Designs
General Linear Models -- #1 things to remember b weight interpretations 1 quantitative predictor 1 quantitative predictor & non-linear component 1 2-group.
Simple Correlation Scatterplots & r Interpreting r Outcomes vs. RH:
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
Design Conditions & Variables Limitations of 2-group designs “Kinds” of Treatment & Control conditions Kinds of Causal Hypotheses Explicating Design Variables.
An Introduction to Classification Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification.
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Calibration & Curve Fitting
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Research Terminology for The Social Sciences.  Data is a collection of observations  Observations have associated attributes  These attributes are.
Moderation & Mediation
More complicated ANOVA models: two-way and repeated measures Chapter 12 Zar Chapter 11 Sokal & Rohlf First, remember your ANOVA basics……….
Statistical Control Statistical Control is about –underlying causation –under-specification –measurement validity –“Correcting” the bivariate relationship.
BIOL 582 Lecture Set 11 Bivariate Data Correlation Regression.
Coding Multiple Category Variables for Inclusion in Multiple Regression More kinds of predictors for our multiple regression models Some review of interpreting.
Curvilinear 2 Modeling Departures from the Straight Line (Curves and Interactions)
Questions to Ask Yourself Regarding ANOVA. History ANOVA is extremely popular in psychological research When experimental approaches to data analysis.
Regression Models w/ 2 Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting the regression.
Mixed-Design ANOVA 5 Nov 2010 CPSY501 Dr. Sean Ho Trinity Western University Please download: treatment5.sav.
Adjusted from slides attributed to Andrew Ainsworth
Regression Models w/ 2-group & Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting.
Plotting Linear Main Effects Models Interpreting 1 st order terms w/ & w/o interactions Coding & centering… gotta? oughta? Plotting single-predictor models.
Regression Models for Quantitative (Numeric) and Qualitative (Categorical) Predictors KNNL – Chapter 8.
Slide 1 Mixed ANOVA (GLM 5) Chapter 15. Slide 2 Mixed ANOVA Mixed: – 1 or more Independent variable uses the same participants – 1 or more Independent.
ANCOVA. What is Analysis of Covariance? When you think of Ancova, you should think of sequential regression, because really that’s all it is Covariate(s)
General Linear Model.
Curvilinear Regression
Meta-Analysis Effect Sizes effect sizes – r, d & OR computing effect sizes estimating effect sizes & other things of which to be careful!
Kxkxk 3-way Factorial Designs kxkxk 3-way designs, effects & causality 3-way interactions describing effects and checking if they are. “descriptive” or.
Handout Twelve: Design & Analysis of Covariance
3-way Designs defining 2-way & 3-way interactions all the effects conditional and unconditional effects orthogonality, collinearity & the like kxkxk kxkxq.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, partial correlations & multiple regression Using model plotting to think about ANCOVA & Statistical control Homogeneity.
ANCOVA.
Mixed-Design ANOVA 13 Nov 2009 CPSY501 Dr. Sean Ho Trinity Western University Please download: treatment5.sav.
Linear Discriminant Function Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification & LDF.
1. Refresher on the general linear model, interactions, and contrasts UCL Linguistics workshop on mixed-effects modelling in R May 2016.
Plotting Non-linear & Complex Main Effects Models
Statistics 200 Lecture #5 Tuesday, September 6, 2016
Interactions and Factorial ANOVA
Data, Univariate Statistics & Statistical Inference
Introduction to the General Linear Model (GLM)
Soc 3306a: ANOVA and Regression Models
Soc 3306a Lecture 11: Multivariate 4
Curvilinear Regression
Presentation transcript:

General Linear Model What is the General Linear Model?? A bit of history & our goals Kinds of variables & effects It’s all about the model (not the data)!! –plotting models & interpreting regression weights –model fit –“expected values” –main/simple/interaction & unconditional/conditional effects Coding & Centering

A common way to characterize the GLM is shown below Multiple Regression quant Criterion quant & binary Preds ANOVA quant DV categorical IVs ANCOVA quant DV categorical IVs quant Covariates

We will use “GLM” to mean a model which has a quantitative criterion/dependent variable and any mix of several kinds of predictors/IVs/covariates/controls… Linear quantitative (raw & “re-centered”) Non-linear quantitative (quadratic, cubic, etc) Binary categorical (raw, unit-, dummy-, effect & comparison-coded) Multiple categorical (raw, unit-, dummy-, effect-, trend- & comparison-coded) Categorical interactions Linear interactions Non-linear interactions

There is a lot of “historical baggage” captured this picture of the GLM!! One way to think about it is that three different models each picked a subset of things in the GLM… for example… Multiple regression is usually a “linear quant main effects model” lots of predictors usually linear-only & no interactions ANOVA is “categorical interaction model” minimum number of IVs always look at interactions & sometimes looks at non-linear ANCOVA is a “select combo of ANOVA & Regression” Minimum categorical IVs & quantitative covariates Looks at interactions among IVs – main effects of covariates

Our goal is to be able to put… any sort of variables predictor, IV, covariate, control, etc linear or non-linear quantitative variables… binary, multiple-category, ordered-category variables… categorical, quantitative, mixed & non-linear interactions coded or re-centered so that the associated regression weights can be plotted to reveal the model give as-direct-as-possible tests of RH: & RQ: using either “regression” “ANOVA/ANCOVA” or “GLM” different analysts/packages call different collections of input and output different things have to know how to control input and understand output some are well-documented & understood… some…

Up until now we have limited the kinds of predictor variables in our models to quantitative and binary (usually coded 1 & 2). In this section we will add several new variable types to our repertoire… dummy & effect codes for binary qualitative variables -- different coding strategies allow us to test specific RH: dummy, effect & comparison coding for k-category variables -- lots of categorical or qualitative variables aren’t binary -- k-group variables can be important predictors, covariates, controls, etc – we have to be able to accommodate them in MR & GLM models -- different coding strategies allow us to test specific RH: What were those kinds of variables again ??

Non-linear relationships -- not all the interesting relationships are “linear” or “just linear” ≠ ≠≠ r = 0.0 for both  no linear relationship. But the plot on the left shows a strong relationship – just not a linear relationship Each has a “linear component“ But not the “same shape of relationship” “not-just linear”

2-way interactions - any mix of binary, k-group & quant predictor vars No Yes practice % correct expert novice # sessions depression Drugs Talk Control 3-way interactions - any mix of binary, k-group & quant predictor vars practice% correct expertnovice unpaid paid # sessions depression +1 std -1 std mean Commitment

non-linear interactions – you can miss interactions if only for linear interactions

Plotting Models!!! So far we have emphasized the clear interpretation of regression weights for each type of predictor. Many of the “more complicated” regression models, such as ANCOVA or those with non-linear or interaction terms are plotted, because, as you know, “a picture is worth a lot of words”. Each term in a multiple regression model has an explicit representation in a regression plot as well as an explicit interpretation – usually with multiple parallel phrasing versions. So, we’ll start simple with simple models and learn the correspondence between: interpretation of each regression model term the graphical representation of that term

Very important things to remember… 1) We plot and interpret the model of the data -- not the data if the model fits the data poorly, then we’re carefully describing and interpretingnonsense 2) The regression weights tell us the “expected values of population parameters” sometimes “expected values” from the model match “descriptive values” from sample, but sometimes not… One you know… descriptive marginal means match the expected marginal means from an orthogonal factorial, but not from a non-orthogonal factorial (which have been corrected for other effects in the model) One you’ll learn… descriptive grand mean across groups will match the expected grand mean only if the groups have equal-n (because the descriptive, but not expected value is influenced by “n”)

One you’ll hear a lot, in every model from here on… 3) The interpretation of regression weights in a main effects model is different than in a model including interactions in a main effects model (without interactions) regression weights reflect “main effects” in a model including interactions regression weights reflect “simple effects”

A bit of new language to go with this… Unconditional effect – the effect is not dependent upon higher-order effects effect in a main effect model (with no interaction terms) effect in a model with non-significant interaction(s) “descriptive” lower-order effect in a model with significant interaction(s), but effect matches corresponding “simple effects” Conditional effect – the effect dependent upon higher-order effects can’t happen in a main effects model can’t happen in a model with non-significant interaction(s) “misleading“ lower-order effect in a model with significant interaction(s) and effect doesn’t match corresponding “simple effects”

Coding & Centering predictors for plotting & interpreting Categorical predictors will be converted to dummy codes comparison/control group coded other group a “target group” of one dummy code, coded 1 Quantitative predictors will be centered, usually to the mean centered = score – mean (like for quadratic terms) so, mean = 0 Why? Mathematically – 0s (as control group & mean) simplify the math & minimize collinearity complications Interpretively – the “controlling for” included in multiple regression weight interpretations is really “controlling for all other variables in the model at the value 0” – “0” as the comparison group & mean will make b interpretations simpler and more meaningful