Chapter 7 Multicollinearity. What is in this Chapter? In Chapter 4 we stated that one of the assumptions in the basic regression model is that the explanatory.

Slides:



Advertisements
Similar presentations
Multivariate Regression
Advertisements

Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Specification Error II
Introduction and Overview
Studenmund(2006): Chapter 8
Correlation and regression
Econ 140 Lecture 151 Multiple Regression Applications Lecture 15.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 13 Additional Topics in Regression Analysis
Chapter 10 Simple Regression.
Statistics for Managers Using Microsoft® Excel 5th Edition
Chapter 5 Heteroskedasticity. What is in this Chapter? How do we detect this problem What are the consequences of this problem? What are the solutions?
Chapter 4 Multiple Regression.
Chapter 9 Simultaneous Equations Models. What is in this Chapter? In Chapter 4 we mentioned that one of the assumptions in the basic regression model.
January 6, morning session 1 Statistics Micro Mini Multiple Regression January 5-9, 2008 Beth Ayers.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 13 Introduction to Linear Regression and Correlation Analysis.
Chapter 7 Multicollinearity.
1 Chapter 9 Variable Selection and Model building Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Linear Regression and Correlation Analysis
Predictive Analysis in Marketing Research
Lecture 11 Multivariate Regression A Case Study. Other topics: Multicollinearity  Assuming that all the regression assumptions hold how good are our.
Topic 3: Regression.
Chapter 9 Multicollinearity
Ekonometrika 1 Ekonomi Pembangunan Universitas Brawijaya.
Stat 112: Lecture 9 Notes Homework 3: Due next Thursday
AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Chapter 13.3 Multicollinearity.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: prediction Original citation: Dougherty, C. (2012) EC220 - Introduction.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
Forecasting Revenue: An Example of Regression Model Building Setting: Possibly a large set of predictor variables used to predict future quarterly revenues.
Ordinary Least Squares
Objectives of Multiple Regression
Regression Method.
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 25 Categorical Explanatory Variables.
© 2004 Prentice-Hall, Inc.Chap 15-1 Basic Business Statistics (9 th Edition) Chapter 15 Multiple Regression Model Building.
Forecasting Revenue: An Example of Regression Model Building Setting: Possibly a large set of predictor variables used to predict future quarterly revenues.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
2 Multicollinearity Presented by: Shahram Arsang Isfahan University of Medical Sciences April 2014.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Correlation & Regression
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
ANOVA, Regression and Multiple Regression March
B AD 6243: Applied Univariate Statistics Multiple Regression Professor Laku Chidambaram Price College of Business University of Oklahoma.
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.
MULTICOLLINEARITY: WHAT HAPPENS IF THE REGRESSORS ARE CORRELATED?
Managerial Economics & Decision Sciences Department introduction  inflated standard deviations  the F  test  business analytics II Developed for ©
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Multivariate Regression
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Fundamentals of regression analysis
Chapter 6: MULTIPLE REGRESSION ANALYSIS
Multiple Regression Models
Multiple Regression Chapter 14.
Regression Forecasting and Model Building
BEC 30325: MANAGERIAL ECONOMICS
Chapter 13 Additional Topics in Regression Analysis
Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences? It may be difficult to separate.
BEC 30325: MANAGERIAL ECONOMICS
Presentation transcript:

Chapter 7 Multicollinearity

What is in this Chapter? In Chapter 4 we stated that one of the assumptions in the basic regression model is that the explanatory variables are not exactly linearly related. If they are, then not all parameters are estimable What we are concerned with in this chapter is the case where the individual parameters are not estimable with sufficient precision (because of high standard errors) This often occurs if the explanatory variables are highly intercorrelated (although this condition is not necessary).

What is in this Chapter? This chapter is very important, because multicollinearity is one of the most misunderstood problems in multiple regression There have been several measures for multicollinearity suggested in the literature (variance-inflation factors VIF, condition numbers CN, etc.) This chapter argues that all these are useless and misleading They all depend on the correlation structure of the explanatory variables only.

What is in this Chapter? It is argued here that this is only one of several factors determining high standard errors High intercorrelations among the explanatory variables are neither necessary nor sufficient to cause the multicollinearity problem The best indicators of the problem are the t-ratios of the individual coefficients.

What is in this Chapter? This chapter also discusses the solutions offered for the multicollinearity problem: – ridge regression –principal component regression –dropping of variables However, they are ad hoc and do not help The only solutions are to get more data or to seek prior information

7.1 Introduction Very often the data we use in multiple regression analysis cannot give decisive answers to the questions we pose This is because the standard errors are very high or the t-ratios are very low The confidence intervals for the parameters of interest are thus very wide This sort of situation occurs when the explanatory variables display little variation and/or high intercorrelations

7.1 Introduction The situation where the explanatory variables are highly intercorrelated is referred to as multicollinearity When the explanatory variables are highly intercorrelated, it becomes difficult to disentangle the separate effects of each of the explanatory variables on the explained variable

7.1 Introduction The practical questions we need to ask are how high these intercorrelations have to be to cause problems in our inference about the individual parameters and what we can do about this problem We argue in the subsequent sections that high intercorrelations among the explanatory variables need not necessarily create a problem and some solutions often suggested for the multicollinearity problem can actually lead us on a wrong track The suggested cures are sometimes worse than the disease

7.2 Some Illustrative Examples

In practice, addition or deletion of observations would produce changes in the variances and covariances Thus one of the consequences of high correlation between x1 and x2 is that the parameter estimates would be very sensitive to the addition or deletion of observations This aspect of multicollinearity can be checked in practice by deleting or adding some observations and examining the sensitivity of the estimates to such perturbations

7.2 Some Illustrative Examples

7.3 Some Measures of Multicollinearity

7.4 Problems with Measuring Multicollinearity

We can summarize the conclusions from this illustrative example as follows: –1. It is difficult to define multicollinearity in terms of the correlations between the explanatory variables because the explanatory variables can be redefined in a number of different ways and these can give drastically different measures of intercorrelations. In some cases, these redefinitions may not make sense, but in the example above involving measured income, permanent income, and transitory income, these redefinitions make sense.

7.4 Problems with Measuring Multicollinearity –2. Just because the explanatory variables are uncorrelated it does not mean that we have no problems with inference. Note that the estimate of a and its standard error are the same in equation (7.5) (with the correlation among the explanatory variables equal to 0.95) and in equation (7.6) (with the explanatory variables uncorrelated). –3. Often, though the individual parameters are not precisely estimable, some linear combinations of the parameters are. For instance, in our example, (α + ß) is estimable with good precision. Sometimes, these linear combinations do not make economic sense. But at other times they do.

7.4 Problems with Measuring Multicollinearity

The example above illustrates four different ways of looking at the multicollinearityproblem: –1. Correlation between the explanatory variables L and Y, which is high. This suggests that the multicollinearity may be serious. However, we explained earlier the fallacy in looking at just the correlation coefficients between the explanatory variables. –2. Standard errors or t-ratios for the estimated coefficients: In this example the t-ratios are significant, suggesting that multicollinearity might not be serious

7.4 Problems with Measuring Multicollinearity –3. Stability of the estimated coefficients when some observations are deleted. Again one might conclude that multicollinearity is not serious, if one uses a 5% level of significance for this test. –4. Examining the predictions from the model: If multicollinearity is a serious problem, the predictions from the model would be worse than those from a model that includes only a subset of the set of explanatory variables The last criterion should be applied if prediction is the object of the analysis. Otherwise,it would be advisable to consider the second and third criteria. The first criterion is not useful, as we have so frequently emphasized

7.5 Solutions to the Multicollinearity Problem: Ridge Regression

7.6 Principal Component Regression

7.7 Dropping Variables

7.8 Miscellaneous Other Solutions Using Ratios or First Differences –We have discussed the method of using ratios in our discussion of heteroskedasticity (Chapter 5) and first differences in our discussion of autocorrelation (Chapter 6) – Although these procedures might reduce the intercorrelations among the explanatory variables, they should be used on the basis of the considerations discussed in those chapters, not as a solution to the collinearity problem

7.8 Miscellaneous Other Solutions Using Extraneous Estimates

7.8 Miscellaneous Other Solutions Getting More Data –One solution to the multicollinearity problem that is often suggested is to "go and get more data." –Actually, the extraneous estimators case we have discussed also falls in this category (we look for another model with common parameters and the associated dataset). –Sometimes using quarterly or monthly data instead of annual data helps us in getting better estimates –However, we might be adding more sources of variation like seasonality. –In any case, since weak data and inadequate information are the sources of our problem, getting more data will help matters.