Non-linear Components for Multiple Regression Why we might need non-linear components Type of non-linear components Squared-term components & their disadvantages.

Slides:



Advertisements
Similar presentations
Questions From Yesterday
Advertisements

How to Factor Quadratics of the Form
Graphing Linear Inequalities in Two Variables
Graphing Techniques and Interpreting Graphs
A4.e How Do I Graph The Solution Set of A Linear Inequality in Two Variables? Course 3 Warm Up Problem of the Day Lesson Presentation.
Dummy Dependent variable Models
Bivariate &/vs. Multivariate
Chapter 4: Basic Estimation Techniques
The Game of Algebra or The Other Side of Arithmetic The Game of Algebra or The Other Side of Arithmetic © 2007 Herbert I. Gross by Herbert I. Gross & Richard.
Your boss asks… How many of these things do we have to sell before we start making money? Use your arrow keys to navigate the slides.
Pricing I: Linear Demand
Chapter 7: The Distribution of Sample Means
Risk and Return Learning Module.
PHYSICS UNIT 1: KINEMATICS (Describing Motion)
Number bonds to 10,
Order of Operations And Real Number Operations
This expression is a statement. What are we going to do today? Today, we: ______________________________ _____________________________. Checking for Understanding.
Transformations & Data Cleaning
Linear Regression.
Research Hypotheses and Multiple Regression: 2 Comparing model performance across populations Comparing model performance across criteria.
Statistical Techniques I EXST7005 Multiple Regression.
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
Chapter 10 Regression. Defining Regression Simple linear regression features one independent variable and one dependent variable, as in correlation the.
Longitudinal Experiments Larry V. Hedges Northwestern University Prepared for the IES Summer Research Training Institute July 28, 2010.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
Multiple Regression Models: Some Details & Surprises Review of raw & standardized models Differences between r, b & β Bivariate & Multivariate patterns.
Intro to Statistics for the Behavioral Sciences PSYC 1900
Introduction the General Linear Model (GLM) l what “model,” “linear” & “general” mean l bivariate, univariate & multivariate GLModels l kinds of variables.
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
An Introduction to Classification Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Chapter 5: z-scores.
Algebra Problems… Solutions
8/10/2015Slide 1 The relationship between two quantitative variables is pictured with a scatterplot. The dependent variable is plotted on the vertical.
Chapter 13: Inference in Regression
Example 11.4 Demand and Cost for Electricity Modeling Possibilities.
Programming with Alice Computing Institute for K-12 Teachers Summer 2011 Workshop.
Regression Examples. Gas Mileage 1993 SOURCES: Consumer Reports: The 1993 Cars - Annual Auto Issue (April 1993), Yonkers, NY: Consumers Union. PACE New.
Copyright © Cengage Learning. All rights reserved. CHAPTER 5 Extending the Number System.
1 Psych 5510/6510 Chapter 10. Interactions and Polynomial Regression: Models with Products of Continuous Predictors Spring, 2009.
Chapter 9 Analyzing Data Multiple Variables. Basic Directions Review page 180 for basic directions on which way to proceed with your analysis Provides.
?v=cqj5Qvxd5MO Linear and Quadratic Functions and Modeling.
Curvilinear 2 Modeling Departures from the Straight Line (Curves and Interactions)
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Introduction to regression 3D. Interpretation, interpolation, and extrapolation.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Regression Models w/ 2 Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting the regression.
Relationships If we are doing a study which involves more than one variable, how can we tell if there is a relationship between two (or more) of the.
Slide 1 The introductory statement in the question indicates: The data set to use (2001WorldFactBook) The task to accomplish (association between variables)
Chapter 8 Linear Regression. Slide 8- 2 Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the.
GRAPHING AND RELATIONSHIPS. GRAPHING AND VARIABLES Identifying Variables A variable is any factor that might affect the behavior of an experimental setup.
Regression Analysis: Part 2 Inference Dummies / Interactions Multicollinearity / Heteroscedasticity Residual Analysis / Outliers.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
Correlation They go together like salt and pepper… like oil and vinegar… like bread and butter… etc.
Answering Descriptive Questions in Multivariate Research When we are studying more than one variable, we are typically asking one (or more) of the following.
General Linear Model What is the General Linear Model?? A bit of history & our goals Kinds of variables & effects It’s all about the model (not the data)!!
Chapter 4 More on Two-Variable Data. Four Corners Play a game of four corners, selecting the corner each time by rolling a die Collect the data in a table.
Kinematics Graphical Analysis of Motion. Goal 2: Build an understanding of linear motion. Objectives – Be able to: 2.04 Using graphical and mathematical.
Slide Copyright © 2009 Pearson Education, Inc. Types of Distributions Rectangular Distribution J-shaped distribution.
Multiple Regression.
Theme 6. Linear regression
QM222 Class 16 & 17 Today’s New topic: Estimating nonlinear relationships QM222 Fall 2017 Section A1.
Multiple Regression.
Introduction to the General Linear Model (GLM)
Multiple Regression.
Properties of the Real Numbers Part I
Scatter Plot 3 Notes 2/6/19.
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Non-linear Components for Multiple Regression Why we might need non-linear components Type of non-linear components Squared-term components & their disadvantages Centered-squared-term components & their advantages Quadratic terms, interactions & regression weights

Non-linear Components Although linear regression models do a very good job of capturing or approximating most of the relationships between quantitative variables that psychologists study, there are several notable exceptions. Study of skills acquisition or other accrual tasks often show... initial learning is slow in earlier trials but then speeds up (a) … learning rates seem to reach a point of diminishing returns (b) … or both (c) … depending upon the novelty of the learning task and the number of trials examined ab c

“Modeling” these non-linear relationships has two major forms… actually fitting the mathematical function of the relationship including nonlinear terms into linear regression literally asking if there is a linear relationship between the criterion and a non-linear transformation of the predictor very similar to the process of “trend analysis” in ANOVA As you might imagine, the latter is more common (and easier) The most common non-linear term (and the only one we’ll cover here) is the quadratic term -- has two forms … U-shaped Inverted-U-shaped

The most common way of including a quadratic term into a linear regression model is to construct a new variable – a “squared term” that is the square of the original variable X  X 2 include the squared term in the model +b means a U-shaped quadratic component -b means an inverted-U-shaped quadratic component determine if the term “adds” to the model t-test of “b” F-test of R²-change w/ & w/o squared term For example, if we wanted to test for a non-linear relationship between practice and performance… compute practsq = pract * pract dep = perform / enter pract paractsq

However, this common method is not doing quite what we had in mind … The practsq term really combines the linear and the quadratic trends into a single term a “quadratic only” trend will have the same DV values at each end of the predictor continuum, with a hump in between however, squaring makes initially larger numbers “more larger” than initially smaller numbers (2*2=4 5*5=25) so, the squaring process alone will give you a “curved” shape (linear + quadratic), but won’t give you a “humped” or “pure quadratic” shape

There are two consequences of working with this “combined” term instead of a “quadratic only” term… the bivariate analysis using this constructed variable won’t be testing what you had intended the multivariate analysis will have increased collinearity - because both the linear and the “quadratic” term have a linear component Remember that, in the multiple regression, the b for the quadratic term will be “controlled for” the linear term, and so will provide a test of the “pure” quadratic trend

Is there a way to get a “pure” quadratic term ??? Sure there is… first center the predictor (subtract the mean of the predictor from each person’s score) X  (X – M) square this centered score (X – M) 2 How does that work ??? Consider two persons with scores equally far above and below the mean (mean = 10, p1 = 8 p2 = 12) After centering, they are equally far from the mean, but with opposite signs (p1 = -2 p2 = 2) When the centered scores are squared, they are now equally far from the mean & iin the same direction (p1 = p2 = 4) If there is a “pure” quadratic term, these squared centered scores are now linearly related to the criterion

Original scores show a “pure” quad trend w/ criterion Centered scores show the same trend w/ criterion Squared centered scores show a linear trend with criterion 0 4 Thus, a quadratic trend between the criterion and the predictor is revealed as a linear trend between the centered & squared predictor and the criterion.

But wait – there’s more!!! How do we interpret the quadratic term regression weight ? Some background…. To answer that, you have to realize that a quadratic term is “a kind of interaction” – HUH?? the quadratic term tells how the slope of the linear y-x relationship changes as x changes. So, it is a conditional statement… So, it is an interaction! +b indicates that the y-x linear slope is getting more positive as x increases -b indicates that the y-x linear slope is getting less positive (more negative) as x increases

Not sure that makes sense??? Take a look at these data… Practice % Correct There is a definite +linear trend! On average, more practice is better! (a main effect) However, there is also a quadratic trend! How much the next practice will help depends on how much practices you’ve already done. (an interaction) 10 practices will help someone who has done 0 practices MORE THAN 10 practices will help someone who has already done 20 !

So, how do we interpret the quadratic term regression weight ? Well, there’s a bit of a problem, which can be expressed 2 ways: First, there’s no way to fix the linear term and allow the quadratic term to vary! Said differently, if X changes X 2 changes…. Second, since it is a kind of interaction term, the change in Y for a 1-unit change in X depends on the starting value of X !!! Here’s an example… We’re going to start with a model that has no linear trend, but a quadratic trend (making it easier to track the quadratic!) Y’ = 0*(X-M) + 2*(X-M)

So, how do we interpret the quadratic term regression weight ? For X=0  X=1 Y’Δ  which is the quadratic coefficient = 2 For X=1  X=2 Y’Δ  ?????

We’re going to think about it as an interaction term… In what direction & by how much does the slope of the Y-X regression line as X increases by 1 ??? b-quad tells how Y-X slope changes as X changes As X increases, the Y-X slope becomes more positive As X decreases, the Y-X slope becomes less positive b-linear tells Y-X slope when x = 0 Y’ = 0*(X-M) + 2*(X-M)