Multiple Regression. From last time There were questions about the bowed shape of the confidence limits around the regression line, both for limits around.

Slides:



Advertisements
Similar presentations
Topic 12: Multiple Linear Regression
Advertisements

Statistical Techniques I EXST7005 Multiple Regression.
Multiple Regression. Outline Purpose and logic : page 3 Purpose and logic : page 3 Parameters estimation : page 9 Parameters estimation : page 9 R-square.
Regression single and multiple. Overview Defined: A model for predicting one variable from other variable(s). Variables:IV(s) is continuous, DV is continuous.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
MULTIPLE REGRESSION. OVERVIEW What Makes it Multiple? What Makes it Multiple? Additional Assumptions Additional Assumptions Methods of Entering Variables.
Psychology 202b Advanced Psychological Statistics, II February 1, 2011.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Chapter 12 Simple Regression
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
Statistical Analysis SC504/HS927 Spring Term 2008 Session 5: Week 20: 15 th February OLS (2): assessing goodness of fit, extension to multiple regression.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
Lecture 6: Multiple Regression
Lecture 24: Thurs., April 8th
Chapter 11 Multiple Regression.
Multiple Linear Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Multiple Regression Dr. Andy Field.
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Multiple Linear Regression Response Variable: Y Explanatory Variables: X 1,...,X k Model (Extension of Simple Regression): E(Y) =  +  1 X 1 +  +  k.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Chapter 13: Inference in Regression
Simple Linear Regression Models
Correlation and Linear Regression. Evaluating Relations Between Interval Level Variables Up to now you have learned to evaluate differences between the.
Introduction to Linear Regression
Extension to Multiple Regression. Simple regression With simple regression, we have a single predictor and outcome, and in general things are straightforward.
1 G Lect 6M Comparing two coefficients within a regression equation Analysis of sets of variables: partitioning the sums of squares Polynomial curve.
1 Psych 5510/6510 Chapter 10. Interactions and Polynomial Regression: Models with Products of Continuous Predictors Spring, 2009.
Chapter 9 Analyzing Data Multiple Variables. Basic Directions Review page 180 for basic directions on which way to proceed with your analysis Provides.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Multiple Linear Regression Partial Regression Coefficients.
Regression Analysis A statistical procedure used to find relations among a set of variables.
Questions to Ask Yourself Regarding ANOVA. History ANOVA is extremely popular in psychological research When experimental approaches to data analysis.
Department of Cognitive Science Michael J. Kalsher Adv. Experimental Methods & Statistics PSYC 4310 / COGS 6310 Regression 1 PSYC 4310/6310 Advanced Experimental.
©2006 Thomson/South-Western 1 Chapter 14 – Multiple Linear Regression Slides prepared by Jeff Heyl Lincoln University ©2006 Thomson/South-Western Concise.
Chapter 16 Data Analysis: Testing for Associations.
Analysis Overheads1 Analyzing Heterogeneous Distributions: Multiple Regression Analysis Analog to the ANOVA is restricted to a single categorical between.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice- Hall, Inc. Chap 14-1 Business Statistics: A Decision-Making Approach 6 th Edition.
Ch15: Multiple Regression 3 Nov 2011 BUSI275 Dr. Sean Ho HW7 due Tues Please download: 17-Hawlins.xls 17-Hawlins.xls.
Outline of Today’s Discussion
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 10 th Edition.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
1 Forecasting/ Causal Model MGS Forecasting Quantitative Causal Model Trend Time series Stationary Trend Trend + Seasonality Qualitative Expert.
Applied Quantitative Analysis and Practices LECTURE#28 By Dr. Osman Sadiq Paracha.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Chapter 8 Relationships Among Variables. Outline What correlational research investigates Understanding the nature of correlation What the coefficient.
Venn diagram shows (R 2 ) the amount of variance in Y that is explained by X. Unexplained Variance in Y. (1-R 2 ) =.36, 36% R 2 =.64 (64%)
Week of March 23 Partial correlations Semipartial correlations
Biostatistics Regression and Correlation Methods Class #10 April 4, 2000.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
Michael J. Kalsher PSYCHOMETRICS MGMT 6971 Regression 1 PSYC 4310 Advanced Experimental Methods and Statistics © 2014, Michael Kalsher.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Stats Methods at IC Lecture 3: Regression.
Multiple Regression.
Correlation, Bivariate Regression, and Multiple Regression
Review Guess the correlation
Regression 11/6.
Regression 10/29.
Multiple Regression Analysis and Model Building
Regression.
Least Squares ANOVA & ANCOV
Multiple Regression – Part II
CHAPTER 29: Multiple Regression*
Multiple Regression.
Multiple Linear Regression
Presentation transcript:

Multiple Regression

From last time There were questions about the bowed shape of the confidence limits around the regression line, both for limits around the mean and the individuals should be curved because they are estimates. In theory if you knew the population values you would not need to bow CLs around the individuals. You just shift the density around following the regession line.

CLs You do not have the same percision across the entire range. The CLs around the mean are based on: The CLs for a new person has the distance from the mean of x in the formula as well. Overall estimated variance in y

Multiple Regression You have seen that you can include polynomials in a regression model. What about including entirely different predictors? Say you want to predict the blood pressure of daft scientists before they give talks at an international conference. You could predict with many single predictors or as a set: –age –the size of the audience –number of “hawt” potential evil lackeys in the front row

Explaining variance Age Audience sizeLackey Quality Total variance in blood pressure

Explained Multivariate Variance The total variance explained depends on how correlated the predictors are. You want to have a global R 2 to indicate the amount of the variance explained by the model and also measures of the contributions of the predictors.

Multicolinearity Even though audience size and lackey quality both allow for you to predict the heart rate of the mad scientists, the amount of variance that is uniquely associated with the lackey variable is very little. When you have very correlated predictors, they can’t uniquely explain variance. You can end up with a model that is statistically significant with a big R 2 but none of the individual predictors is statistically significant.

Stop it before it starts Before you do a multiple regression use subject matter knowledge to remove highly correlated predictors. Look at the bivaraite correlation coefficients between the predictors. If you have highly correlated variables use subject matter and pragmatism to decide which varible to put in the model.

Partitioning Variance What is the unique contribution of each variable? There are different formulas for adding up the sum of squares SS (in the variance). Sequential (aka type 1 SS) lets the first variable explain as much variance as it can then add in the second and see if it can explain any of the remaining variance. Simultaneous (aka type 2 SS) put the first all the variables in at the same time and let them divvy up the common variance. Simultaneous with interactions (aka type 3 SS) have the variable try to explain all they can and consider they are used interactions. Type 4 SS makes my head hurt.

SAS vs. R S-Plus and SAS use the same formulas for SS but R does not use the same formula for Type 3 SS (and I have never tested it on Type 4 SS).

Partial and Semipartial Correlations If you want to look at the correlation between a predictor (a) and an outcome (z) controlling for the impact of a second predictor (b) you can do a partial correlation to remove the impact b on both. You can also remove the correlation between b and a only you can do a semi-partial correlation.

Hierarchical Stepwise Regression If you take 2 nd and 3 rd quarter statistics you will learn the details on how to compare two models. Hierarchical stepwise regression is the process of figuring out what variables matter (in advance) and adding them to a model to see if you improve the quality of the model as you add them. People frequently look at the R 2 for the models and/or use AIC. This is a fine thing to do so long as you keep track of the comparison you make and report it.

Automatic Stepwise Regression These are BAD BAD BAD. You feed the software a set of variables and tell it put them into the model one at a time to find the predictor that explains the most outcome variance. Once that is put into the model, add all the remaining ones one at a time to see if the residual variance is reduced with the second variable. Repeat over and over. Some of the methods subtract variables instead of adding (others do both). The Type 1 error is astronomical. These methods have horrible properties. Adding in completely random variables affects the model.