 Population multiple regression model  Data for multiple regression  Multiple linear regression model  Confidence intervals and significance tests.

Slides:



Advertisements
Similar presentations
Topic 12: Multiple Linear Regression
Advertisements

Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Inference for Regression
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Chapter 12 Simple Linear Regression
Chapter 10 Simple Regression.
Chapter 12 Simple Regression
Chapter 13 Introduction to Linear Regression and Correlation Analysis
1 Pertemuan 13 Uji Koefisien Korelasi dan Regresi Matakuliah: A0392 – Statistik Ekonomi Tahun: 2006.
Pengujian Parameter Koefisien Korelasi Pertemuan 04 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
Chapter Topics Types of Regression Models
Chapter 11 Multiple Regression.
Introduction to Probability and Statistics Linear Regression and Correlation.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Simple Linear Regression and Correlation
Chapter 7 Forecasting with Simple Regression
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Introduction to Linear Regression and Correlation Analysis
Chapter 13: Inference in Regression
© 2002 Prentice-Hall, Inc.Chap 14-1 Introduction to Multiple Regression Model.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 1 Slide © 2003 Thomson/South-Western Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved OPIM 303-Lecture #9 Jose M. Cruz Assistant Professor.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
CHAPTER 14 MULTIPLE REGRESSION
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
MBP1010H – Lecture 4: March 26, Multiple regression 2.Survival analysis Reading: Introduction to the Practice of Statistics: Chapters 2, 10 and 11.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
Simple Linear Regression ANOVA for regression (10.2)
Chapter 13 Multiple Regression
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Lecture 10: Correlation and Regression Model.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Chapter 10 Inference for Regression
ANOVA, Regression and Multiple Regression March
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
INTRODUCTION TO MULTIPLE REGRESSION MULTIPLE REGRESSION MODEL 11.2 MULTIPLE COEFFICIENT OF DETERMINATION 11.3 MODEL ASSUMPTIONS 11.4 TEST OF SIGNIFICANCE.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Chapter 12: Correlation and Linear Regression 1.
Stats Methods at IC Lecture 3: Regression.
Chapter 13 Simple Linear Regression
The simple linear regression model and parameter estimation
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
Statistics for Managers using Microsoft Excel 3rd Edition
Chapter 11 Simple Regression
Relationship with one independent variable
Chapter 13 Simple Linear Regression
CHAPTER 29: Multiple Regression*
Simple Linear Regression
Relationship with one independent variable
Simple Linear Regression
Presentation transcript:

 Population multiple regression model  Data for multiple regression  Multiple linear regression model  Confidence intervals and significance tests  Squared multiple correlation R 2 1

 Extension of SLR  Statistical model  Estimation of the parameters and interpretation  R-square with MLR  Anova Table  F-Test and t-tests

Most things are conceptually similar to SLR and an extension of what we learned thru chapters 2 and 10. However, most things get much more complex, including the SAS output and learning to interpret it. Lastly, whereas before there usually was a set procedure to analyze the data, we now will have to be more flexible and take things as they come, so to speak.

Population Multiple Regression Equation Up to this point, we have considered in detail the linear regression model in which the mean response, μ y, is related to one explanatory variable x: Usually, more complex linear models are needed in practical situations. There are many problems in which a knowledge of more than one explanatory variable is necessary in order to obtain a better understanding and better prediction of a particular response. 4 In multiple regression, the response variable y depends on p explanatory variables :

Data for Multiple Regression The data for a simple linear regression problem consist of n observations of two variables. Data for multiple linear regression consist of the value of a response variable y and p explanatory variables on each of n cases. We write the data and enter them into software in the form: Variables Casex1x1 x2x2 …xpxp y 1x 11 x 12 …x1px1p y1y1 2x 21 x 22 …x2px2p y2y2 ……………… nxn1xn1 xn2xn2 …x np ynyn 5

 We are interested in finding variables to predict college GPA.  Grades from high school will be used as potential explanatory variables (also called predictors) namely: HSM (math grades), HSS (science grades), and HSE (english grades).  Since there are several explanatory variables or x’s, they need to be distinguished using subscripts: ◦ X 1 =HSM ◦ X 2 =HSS ◦ X 3 =HSE

 Why not do several Simple Linear Regressions ◦ Do GPA with HSM  Significant? ◦ Do GPA with HSS  Significant? ◦ Do GPA with HSE  Significant?  Why not? ◦ Each alone may not explain GPA very well at all but used together they may explain GPA quite well. ◦ Predictors could (and usually do) overlap some, so we’d like to distinguish this overlap (and remove it) if possible.

 Unfortunately because scatterplots are restricted to only 2 axes (Y-axis and X-axis), they are less useful here.  Could plot Y with each predictor separately, like an SLR, but this is just a preliminary look at each of the variables and cannot tell us whether we have a good MLR or not.

 The deviations ε i are assumed to be independent and N(0,  ).  The parameters of the model are:  0,  1,  2,  3, and . ◦ Estimates then become b 0, b 1, b 2, b 3, and

 b 0 is still the intercept  b 1 is the estimated “slope” for  1, it explains how y changes as x 1 changes  Then b 2 is the estimated “slope” for  2, it explains how y changes as x 2 changes ◦ Suppose b 2 = 0.7, then if I change x 2 by 1 point, y changes by 0.7, etc ◦ The exact same interpretation as in SLR

 Predicted values ◦ Given values for x 1, x 2, and x 3, plug those into the regression equation and get a  Residuals ◦ Still Observed – Predicted = y – ◦ Calculations and interpretations are the same  Assumptions ◦ Independence, Linearity, Constant Variance and Normality ◦ Use the same plots, same interpretation

 Confidence intervals for the slopes ◦ Still of the form ◦ *CHANGE!!!  DF = n – p – 1  p is the number of predictors in our model  Recall in SLR we only had 1 predictor, or one x  So, df = n – 1 – 1 = n – 2 for SLR  Now we have p predictors,  For GPA example, df = n – 3 – 1 = n – 4

 Since there is more than 1 predictor, a simple t- test will not suffice to test whether there is a significant linear relationship or not.  The good news… ◦ The fundamental principle is still the same ◦ To help with understanding let’s look at what R-square means…

 Still trying to explain the changes in Y  R-square measures the % of explained variation by the regression line. ◦ So in SLR, this is just the percent explained by the changes in x.  In MLR, it represents the percent explained by all predictors combined simultaneously. ◦ Problem: What if the predictors are overlapping? ◦ In fact, they almost always overlap at least a little bit

 Rectangle represents total variation of Y;  Ovals represent variables; Note OVERLAP! X1 X2 X3 Total Variation of Y

 First, we need a number to describe the total variation (the yellow box) ◦ SST = Total Sums of Squares  Next we need to describe the parts explained by the different predictors. ◦ Unfortunately, for now, all we get is one number for all the variables together. ◦ SSM = Model Sums of Squares (Regression)  Then naturally, R 2 = SSM/SST ◦ The amount of variation the regression explains out of the total variation

 Using the same principle, a single t-test for each predictor is not good enough, we need a collective test for all predictors at the same time. ◦  ANOVA Table

 Breaks up the different pieces of sums of squares ◦ SST = Total variation ◦ SSM = Part explained by the model(regression) ◦ SSE = Leftover unexplained portion  Called Error Sums of Squares  Let’s look again…

ANOVA Table for Multiple Regression SourcedfSum of squares SS Mean square MS FP-value Modelp (from data) MSM=SSM/DFMMSM/MSEFrom Table Errorn − p − 1 (from data) MSE=SSE/DFE Totaln − 1 (from data) 20 SSM = model sums of squares SSE = error sums of squares SST = total sums of squares SST = SSM + SSE DFM = p DFE = n – p – 1 DFT = n – 1 DFT = DFM + DFE

 Additionally, the ANOVA Table tests whether or not there is a significant multiple linear regression ◦ Test statistic is F = MSM/MSE  Under H 0, F has an F distribution (see Table E) with p and n-p-1degrees of freedom (two types): ◦ “Degrees of freedom in the numerator"  DFM = p ◦ “Degrees of freedom in the denominator"  DFE = n – p – 1

P-value

 The hypotheses for the F-test are as follows: H 0 :  1 =  2 =  3 = 0 H α : Some  i ≠ 0 (only need one non-zero  i )  So a rejection of the null indicates that collectively the Xs do well at explaining Y  What it doesn’t show is which of the Xs are doing “the explaining” ◦ We’ll come back to this later

 Since the P-value for the F-test is small, <0.0001, we reject H 0  There is a significant Multiple Linear Regression between Y and the Xs. ◦  Model is useful in predicting Y  My data provides evidence that there is a significant linear regression between GPA and the predictors HSM, HSS, and HSE

 The t-tests now become useful in determining which predictors are actually contributing to the explanation of Y.  There are several different methods of determining which Xs are the best ◦ All possible models selection ◦ Forward selection ◦ Stepwise selection ◦ Backward elimination  We will just learn backward elimination…

 So suppose X1 does a good job explaining Y by itself. Then maybe X2 and X3 are “piggybacking” in to the model. ◦ They themselves aren’t good by themselves but combined with X1, all three look good collectively in the MLR. X1 X2 X3 Total Variation of Y

 A t-test in MLR is similar to what it was in SLR Hypotheses: H 0 :  1 = 0 vs. H a :  1 ≠ 0  The difference is this is testing the usefulness or significance of X1 AFTER X2 and X3 are already in the model.  Added last

P-value very significant HSM significant HSS, HSE not 30

 So both X2 and X3 aren’t significant added last ◦ The backward elimination procedure removes ONLY the single worst predictor, then reruns the MLR with all remaining variables  NOTE: this changes the entire MLR model  Since X2 is the least significant added last, it is removed…  What will the new model be without X2?

 Changes in MLR if the model changes: ◦ The MLR regression line ◦ Parameter estimates ◦ Predicted values, Residuals ◦ R-square ◦ ANOVA table ◦ F-test, T-tests ◦ Assumptions ◦ EVERYTHING!!!

So what’s the next step of backward elimination? ?

 The T-test for X3 now has a P-value better than before, , but should it be removed also?  What are reasonable levels for alpha in MLR? ◦ There is no default alpha level like in SLR. ◦ It just depends on the researcher. ◦ SAS defaults to α = 0.15, why?  Suppose we decide to remove X3 based on this P-value, what will the new model be without X3?

X1 X2 X3 Total Variation of Y Take out X2, what happens in the picture? Then take out X3, what happens?

 Remember what made the regression line in SLR best?  The Least Squares Regression refers to making the ERROR sums of squares as small as possible ◦ If SSE is as small as possible then SSM (the explained variation) is as LARGE as possible!

 The actual data will not fit the regression line exactly: DATA = FIT + RESIDUAL ◦ FIT is the MLR regression line ◦ RESIDUAL (“noise”) =  ◦ The deviations ε i are still assumed to be independent and N(0,  ).

 All the same things for SLR now apply ◦ Can do confidence intervals for slope estimates ◦ R-square ◦ Predictions and Residuals ◦ Prediction Intervals for individuals ◦ Confidence interval for Mean response for a group  Must check model assumptions, if something is violated needs to be addressed. ◦ Interpretation is the same as before.

 Did we get all significant predictors?  Yes! According to the original model containing X1, X2, and X3, we chose the best predictors.  Did we get all significant predictors? ◦ No way! We could have left out predictors to begin with.

R-square helps us see how much room for improvement there is! P-value very significant R 2 is fairly small (20.46%) There could be other variables that can explain the remaining 79.54%

 Book continues the GPA example with a look at adding several more potential predictors (scores from the SAT exam)  So are we done?