Part I – MULTIVARIATE ANALYSIS C3 Multiple Linear Regression II © Angel A. Juan & Carles Serrat - UPC 2007/2008.

Slides:



Advertisements
Similar presentations
All Possible Regressions and Statistics for Comparing Models
Advertisements

Chapter 5 Multiple Linear Regression
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 14-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
Ch11 Curve Fitting Dr. Deshi Ye
1 1 Slide 統計學 Spring 2004 授課教師:統計系余清祥 日期: 2004 年 5 月 11 日 第十二週:建立迴歸模型.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Guide to Using Minitab For Basic Statistical Applications To Accompany Business Statistics: A Decision Making Approach, 6th Ed. Chapter 14: Multiple Regression.
Chapter 13 Multiple Regression
Statistics for Managers Using Microsoft® Excel 5th Edition
MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.
Part I – MULTIVARIATE ANALYSIS
Statistics for Managers Using Microsoft® Excel 5th Edition
Chapter 12 Multiple Regression
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
Part I – MULTIVARIATE ANALYSIS C2 Multiple Linear Regression I
Multiple Regression MARE 250 Dr. Jason Turner.
Lecture 11 Multivariate Regression A Case Study. Other topics: Multicollinearity  Assuming that all the regression assumptions hold how good are our.
Multiple Linear Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Chapter 15: Model Building
Regression Model Building Setting: Possibly a large set of predictor variables (including interactions). Goal: Fit a parsimonious model that explains variation.
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
© 2004 Prentice-Hall, Inc.Chap 15-1 Basic Business Statistics (9 th Edition) Chapter 15 Multiple Regression Model Building.
STA302/ week 111 Multicollinearity Multicollinearity occurs when explanatory variables are highly correlated, in which case, it is difficult or impossible.
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
Business Statistics, 4e, by Ken Black. © 2003 John Wiley & Sons Business Statistics, 4e by Ken Black Chapter 15 Building Multiple Regression Models.
Shonda Kuiper Grinnell College. Statistical techniques taught in introductory statistics courses typically have one response variable and one explanatory.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.
Fitting Curves to Data 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 5: Fitting Curves to Data Terry Dielman Applied Regression.
1 1 Slide © 2005 Thomson/South-Western AK/ECON 3480 M & N WINTER 2006 n Power Point Presentation n Professor Ying Kong School of Analytic Studies and Information.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
1 Experimental Statistics - week 12 Chapter 12: Multiple Regression Chapter 13: Variable Selection Model Checking.
Multiple Logistic Regression STAT E-150 Statistical Methods.
Lecture 3 Introduction to Multiple Regression Business and Economic Forecasting.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Variable Selection 1 Chapter 8 Variable Selection Terry Dielman Applied Regression Analysis:
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Correlation & Regression Analysis
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Specification: Choosing the Independent.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 10 th Edition.
Chap 13-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 13 Multiple Regression and.
1 Building the Regression Model –I Selection and Validation KNN Ch. 9 (pp )
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Using SPSS Note: The use of another statistical package such as Minitab is similar to using SPSS.
Variable selection and model building Part I. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Venn diagram shows (R 2 ) the amount of variance in Y that is explained by X. Unexplained Variance in Y. (1-R 2 ) =.36, 36% R 2 =.64 (64%)
2011 Data Mining Industrial & Information Systems Engineering Pilsung Kang Industrial & Information Systems Engineering Seoul National University of Science.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Stats Methods at IC Lecture 3: Regression.
Chapter 15 Multiple Regression Model Building
Chapter 9 Multiple Linear Regression
John Loucks St. Edward’s University . SLIDES . BY.
Business Statistics, 4e by Ken Black
Business Statistics Multiple Regression This lecture flows well with
Regression Analysis Week 4.
Prepared by Lee Revere and John Large
Product moment correlation
Regression Forecasting and Model Building
Lecture 20 Last Lecture: Effect of adding or deleting a variable
Business Statistics, 4e by Ken Black
Multicollinearity Multicollinearity occurs when explanatory variables are highly correlated, in which case, it is difficult or impossible to measure their.
Presentation transcript:

Part I – MULTIVARIATE ANALYSIS C3 Multiple Linear Regression II © Angel A. Juan & Carles Serrat - UPC 2007/2008

1.3.1: The General Linear Model  Model building is the process of developing an estimated regression equation that describes the relationship between a dependent variable and one or more predictors.  Two major issues in model building are: a)To find the proper functional form (linear, quadratic, etc.) of the relationship b)To select the predictors to be included in the model  Suppose we have data for one response y and k predictors x 1, x 2, …, x k. The General Linear Model (GLM) involving p predictors is:  Note: The regression techniques are not limited to linear relationships. The word “linear” in the term GLM refers only to the fact that β 0, β 1,…, β p all have exponents of one; it does not imply that the relationship between y and the x i ’s is linear. Multicollinearity in a regression model results from the common mistake of putting too many predictors into the model. Such a model is said to be “over-fit”. Inevitably many of the predictors will have effects that are too correlated and cannot be separately estimated

1.3.2: Modeling Curvilinear Relationships  Curvilinear relationships in RA can be handled by using powers of predictors.  File: SALESPEOPLE.MTW Although the output shows that a linear relationship explains a high percentage of the variability in sales, the standardized residual plot suggest that a curvilinear relationship is needed. This standardized residual plot shows that the previous curvilinear pattern has been removed. We want to investigate the relationship between months of employment of the salespeople and the number of units sold.

1.3.3: Interaction between Predictors  When interaction between two predictors is present, we cannot study the effect of one predictor on the response y independently of the other predictor.  File: SALESPRICE.MTW We want to investigate the relationship between price, advertising expenditure and the number of units sold of a product. Note that, at higher selling prices, the effect of increased advertising expenditure diminishes. This fact suggest existence of interaction between the price and advertising predictors. We will use the predictor price*advertising ( P*A ) to account for the effect of interaction. Since the p-value corresponding to the t test for P*A is 0.000, we conclude that interaction is significant, i.e.: the effect of advertising expenditure on sales depends on the price

1.3.4: Transformations of the Response  Often the problem of non-constant variance can be corrected by transforming the response variable to a different scale using a logarithmic transformation, Log(y), or a reciprocal transformation,1/y.  File: KMWEIGHT.MTW We want to investigate the relationship between Km per liter and the vehicle weight. This standardized residual plot shows that the variability in the residuals appears to increase as the fitted value increases. Furthermore, there is a large standardized residual. The wedge-shaped pattern has now disappeared. Moreover, there is not any large standardized residual.

1.3.5: “Linearizing” Nonlinear Models  Models in which the β parameters have exponents other than one are called nonlinear models.  In some cases we can easily transform the nonlinear model into a linear one, as in the case of the exponential model:  Many nonlinear models cannot be transformed into an equivalent linear model. However, such models have had limited use in practical applications.

1.3.6: Model Building: Initial Steps  File: MODELBUILDING.MTW The Correlation Matrix provides the sample correlation coefficients between each pair of variables and the p-value for the corresponding test on significant correlation. The best predictors for Sales seem to be: Accts (R 2 =57%), Time, Poten and Adv. There is multicollinearity between Time and Accts ; hence if Accts were used as a predictor, Time would not add much more explanatory power to the model. A similar problem occurs between Change and Rating. We want to investigate the relationship between Sales and the eight predictors.

1.3.7: Model Building: Predictors Selection  Objective: to select those predictors that provide the “best” regression model.  Alternative methods for selecting the predictors: 1.Stepwise (forward and backward) regression: At each step, a single variable can be added or deleted. The process continues until no more improvement can be obtained by adding or deleting a variable. 2.Forward selection: At each step, a single variable can be added, but variables are never deleted. The process continues until no more improvement can be obtained by adding a variable. 3.Backward elimination: The procedure starts with a model involving all the possible predictors. At each step a variable is deleted. The procedure stops when no more improvement can be obtained by deleting a variable. 4.Best-subsets regression: This is not a one-variable- at-a-time method. It evaluates regression models involving different subsets of the predictors. Iterative methods The stopping criterion is based on the F statistic They provide a “good” regression model with little multicollinearity problems, but not necessarily the “best” model (the one with the highest R 2 ) Iterative methods The stopping criterion is based on the F statistic They provide a “good” regression model with little multicollinearity problems, but not necessarily the “best” model (the one with the highest R 2 ) It provides the “best” regression model for the given data. Functions of the predictors (e.g.: z = x 1 * x 2 ) can be used to create new predictors for use with any of the methods presented here.

1.3.8: Model Building: Stepwise Regression  File: MODELBUILDING.MTW  Stat > Regression > Stepwise... In this example, Stepwise Regression takes 5 steps. At the end of the procedure, five predictors ( Accts, Adv, Poten, Share and Change ) have been selected for the regression model. Note that R-Sq(adj) = 88,94% for the last model. In Stepwise (Forward and Backward) Regression, a significance level of α=0.15 is recommended both for add or delete new variables to/from the model. At each step: 1. The already-in-the-model predictor with the highest non-significant p- value (p-value > α) will be deleted from the model 2. The not-already-in-the-model predictor with the lowest significant p-value (p-value <= α) will enter the model 3. P-values are recalculated At each step: 1. The already-in-the-model predictor with the highest non-significant p- value (p-value > α) will be deleted from the model 2. The not-already-in-the-model predictor with the lowest significant p-value (p-value <= α) will enter the model 3. P-values are recalculated

1.3.9: Model Building: Forward Selection  File: MODELBUILDING.MTW  Stat > Regression > Stepwise... In this example, Forward Selection takes 6 steps. At the end of the procedure, six predictors ( Accts, Adv, Poten, Share, Change and Time ) have been selected for the regression model. Note that R-Sq(adj) = 89,38% for the last model. In Forward Selection, a significance level of α=0.25 is recommended for add new variables to the model. Mallows’ C-p statistic represents a subsetting criterion to be used in selecting a reduced model without multicollinearity problems. A rule of thumb is to select a model in which the value of Cp is close to the number of terms, including any constant term, in the model. At each step: 1. The not-already-in-the-model predictor with the lowest significant p-value (p-value <= α) will enter the model 2. P-values are recalculated At each step: 1. The not-already-in-the-model predictor with the lowest significant p-value (p-value <= α) will enter the model 2. P-values are recalculated

1.3.10: Model Building: Backward Eliminat.  File: MODELBUILDING.MTW  Stat > Regression > Stepwise... In this example, Backward Elimination takes 4 steps. At the end of the procedure, five predictors ( Time, Poten, Adv, Share, and Change ) have been selected for the regression model. Note that R-Sq(adj) = 89,27% for the last model. In Backward Elimination, a significance level of α=0.1 is recommended for delete variables from the model. At each step: 1. The already-in-the-model predictor with the highest non- significant p-value (p-value > α) will be deleted from the model 2. P-values are recalculated At each step: 1. The already-in-the-model predictor with the highest non- significant p-value (p-value > α) will be deleted from the model 2. P-values are recalculated

1.3.11: Model Building: Best-Subsets Reg.  File: MODELBUILDING.MTW  Stat > Regression > Best Subsets... In this example, the adjusted coefficient of determination is largest for the model with six predictors ( Time, Poten, Adv, Share, Change and Accts ). However, the best model with four independent variables ( Poten, Adv, Share and Accts ) has an adjusted coefficient of determination almost as high. The Best-Subsets output identifies the two best one-predictor models, the two best two-predictor models, and so on. The criterion used which models are best for any number of predictors is the value of R 2. All other things being equal, a simpler model with fewer variables is usually preferred.