Introduction and Overview

Slides:



Advertisements
Similar presentations
Managerial Economics in a Global Economy
Advertisements

Welcome to Econ 420 Applied Regression Analysis
The Multiple Regression Model.
Random Assignment Experiments
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Specifying an Econometric Equation and Specification Error
Specification Error II
Studenmund(2006): Chapter 8
Multicollinearity Multicollinearity - violation of the assumption that no independent variable is a perfect linear function of one or more other independent.
The Use and Interpretation of the Constant Term
Choosing a Functional Form
Multiple Linear Regression Model
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Chapter 7 Multicollinearity. What is in this Chapter? In Chapter 4 we stated that one of the assumptions in the basic regression model is that the explanatory.
Multicollinearity Omitted Variables Bias is a problem when the omitted variable is an explanator of Y and correlated with X1 Including the omitted variable.
Further Inference in the Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Topic 3: Regression.
Multiple Regression and Correlation Analysis
Ekonometrika 1 Ekonomi Pembangunan Universitas Brawijaya.
Stat 112: Lecture 9 Notes Homework 3: Due next Thursday
AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Chapter 13.3 Multicollinearity.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
Forecasting Revenue: An Example of Regression Model Building Setting: Possibly a large set of predictor variables used to predict future quarterly revenues.
Ordinary Least Squares
What Is Hypothesis Testing?
Objectives of Multiple Regression
Chapter 15 Forecasting Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Chapter 11 Simple Regression
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
  What is Econometrics? Econometrics literally means “economic measurement” It is the quantitative measurement and analysis of actual economic and business.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Correlation.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Section 10-1 Review and Preview.
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
MultiCollinearity. The Nature of the Problem OLS requires that the explanatory variables are independent of error term But they may not always be independent.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Multiple Collinearity, Serial Correlation,
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Twelve.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Pure Serial Correlation
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Stat 112 Notes 9 Today: –Multicollinearity (Chapter 4.6) –Multiple regression and causal inference.
Chapter 2 Ordinary Least Squares Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
5-1 MGMG 522 : Session #5 Multicollinearity (Ch. 8)
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-1 Basic Mathematical tools Today, we will review some basic mathematical tools. Then we.
MULTICOLLINEARITY: WHAT HAPPENS IF THE REGRESSORS ARE CORRELATED?
Basic Estimation Techniques
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Fundamentals of regression analysis
Pure Serial Correlation
Serial Correlation and Heteroskedasticity in Time Series Regressions
Chapter 6: MULTIPLE REGRESSION ANALYSIS
MULTICOLLINEARITY: WHAT HAPPENS IF THE REGRESSORS ARE CORRELATED
I271b Quantitative Methods
Serial Correlation and Heteroscedasticity in
MULTICOLLINEARITY: WHAT HAPPENS IF THE REGRESSORS ARE CORRELATED
Chapter 7: The Normality Assumption and Inference with OLS
Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences? It may be difficult to separate.
Financial Econometrics Fin. 505
Financial Econometrics Fin. 505
Serial Correlation and Heteroscedasticity in
Presentation transcript:

Introduction and Overview The next three chapters deal with violations of the Classical Assumptions and remedies for those violations This chapter addresses multicollinearity; the next two chapters are on serial correlation and heteroskedasticity For each of these three problems, we will attempt to answer the following questions: 1. What is the nature of the problem? 2. What are the consequences of the problem? 3. How is the problem diagnosed? 4. What remedies for the problem are available? © 2011 Pearson Addison-Wesley. All rights reserved. 1

Perfect Multicollinearity Perfect multicollinearity violates Classical Assumption VI, which specifies that no explanatory variable is a perfect linear function of any other explanatory variables The word perfect in this context implies that the variation in one explanatory variable can be completely explained by movements in another explanatory variable A special case is that of a dominant variable: an explanatory variable is definitionally related to the dependent variable An example would be (Notice: no error term!): X1i = α0 + α1X2i (8.1) where the αs are constants and the Xs are independent variables in: Yi = β0 + β1X1i + β2X2i + εi (8.2) Figure 8.1 illustrates this case © 2011 Pearson Addison-Wesley. All rights reserved. 2

Figure 8.1 Perfect Multicollinearity © 2011 Pearson Addison-Wesley. All rights reserved. 3

Perfect Multicollinearity (cont.) What happens to the estimation of an econometric equation where there is perfect multicollinearity? OLS is incapable of generating estimates of the regression coefficients most OLS computer programs will print out an error message in such a situation What is going on? Essentially, perfect multicollinearity ruins our ability to estimate the coefficients because the perfectly collinear variables cannot be distinguished from each other: You cannot “hold all the other independent variables in the equation constant” if every time one variable changes, another changes in an identical manner! Solution: one of the collinear variables must be dropped (they are essentially identical, anyway) © 2011 Pearson Addison-Wesley. All rights reserved. 4

Imperfect Multicollinearity Imperfect multicollinearity occurs when two (or more) explanatory variables are imperfectly linearly related, as in: X1i = α0 + α1X2i + ui (8.7) Compare Equation 8.7 to Equation 8.1 Notice that Equation 8.7 includes ui, a stochastic error term This case is illustrated in Figure 8.2 © 2011 Pearson Addison-Wesley. All rights reserved. 5

Figure 8.2 Imperfect Multicollinearity © 2011 Pearson Addison-Wesley. All rights reserved. 6

The Consequences of Multicollinearity There are five major consequences of multicollinearity: 1. Estimates will remain unbiased 2. The variances and standard errors of the estimates will increase: a. Harder to distinguish the effect of one variable from the effect of another, so much more likely to make large errors in estimating the βs than without multicollinearity b. As a result, the estimated coefficients, although still unbiased, now come from distributions with much larger variances and, therefore, larger standard errors (this point is illustrated in Figure 8.3) © 2011 Pearson Addison-Wesley. All rights reserved. 7

Figure 8.3 Severe Multicollinearity Increases the Variances of the s © 2011 Pearson Addison-Wesley. All rights reserved. 8

The Consequences of Multicollinearity (cont.) 3. The computed t-scores will fall: a. Recalling Equation 5.2, this is a direct consequence of 2. above 4. Estimates will become very sensitive to changes in specification: a. The addition or deletion of an explanatory variable or of a few observations will often cause major changes in the values of the s when significant multicollinearity exists b. For example, if you drop a variable, even one that appears to be statistically insignificant, the coefficients of the remaining variables in the equation sometimes will change dramatically c. This is again because with multicollinearity, it is much harder to distinguish the effect of one variable from the effect of another 5. The overall fit of the equation and the estimation of the coefficients of nonmulticollinear variables will be largely unaffected © 2011 Pearson Addison-Wesley. All rights reserved. 9

The Detection of Multicollinearity First realize that that some multicollinearity exists in every equation: all variables are correlated to some degree (even if completely at random) So it’s really a question of how much multicollinearity exists in an equation, rather than whether any multicollinearity exists There are basically two characteristics that help detect the degree of multicollinearity for a given application: 1. High simple correlation coefficients 2. High Variance Inflation Factors (VIFs) We will now go through each of these in turn: © 2011 Pearson Addison-Wesley. All rights reserved. 10

High Simple Correlation Coefficients If a simple correlation coefficient, r, between any two explanatory variables is high in absolute value, these two particular Xs are highly correlated and multicollinearity is a potential problem How high is high? Some researchers pick an arbitrary number, such as 0.80 A better answer might be that r is high if it causes unacceptably large variances in the coefficient estimates in which we’re interested. Caution in case of more than two explanatory variables: Groups of independent variables, acting together, may cause multicollinearity without any single simple correlation coefficient being high enough to indicate that multicollinearity is present As a result, simple correlation coefficients must be considered to be sufficient but not necessary tests for multicollinearity © 2011 Pearson Addison-Wesley. All rights reserved. 11

High Variance Inflation Factors (VIFs) The variance inflation factor (VIF) is calculated from two steps: 1. Run an OLS regression that has Xi as a function of all the other explanatory variables in the equation—For i = 1, this equation would be: X1 = α1 + α2X2 + α3X3 + … + αKXK + v (8.15) where v is a classical stochastic error term Calculate the variance inflation factor for : (8.16) where is the unadjusted from step one © 2011 Pearson Addison-Wesley. All rights reserved. 12

High Variance Inflation Factors (VIFs) (cont.) From Equation 8.16, the higher the VIF, the more severe the effects of mulitcollinearity How high is high? While there is no table of formal critical VIF values, a common rule of thumb is that if a given VIF is greater than 5, the multicollinearity is severe As the number of independent variables increases, it makes sense to increase this number slightly Note that the authors replace the VIF with its reciprocal, , called tolerance, or TOL Problems with VIF: No hard and fast VIF decision rule There can still be severe multicollinearity even with small VIFs VIF is a sufficient, not necessary, test for multicollinearity © 2011 Pearson Addison-Wesley. All rights reserved. 13

Remedies for Multicollinearity Essentially three remedies for multicollinearity: Do nothing: a. Multicollinearity will not necessarily reduce the t-scores enough to make them statistically insignificant and/or change the estimated coefficients to make them differ from expectations b. the deletion of a multicollinear variable that belongs in an equation will cause specification bias 2. Drop a redundant variable: a. Viable strategy when two variables measure essentially the same thing b. Always use theory as the basis for this decision! © 2011 Pearson Addison-Wesley. All rights reserved. 14

Remedies for Multicollinearity (cont.) Increase the sample size: This is frequently impossible but a useful alternative to be considered if feasible The idea is that the larger sample normally will reduce the variance of the estimated coefficients, diminishing the impact of the multicollinearity © 2011 Pearson Addison-Wesley. All rights reserved. 15

Table 8.1a © 2011 Pearson Addison-Wesley. All rights reserved. 16

Table 8.1a © 2011 Pearson Addison-Wesley. All rights reserved. 17

Table 8.2a © 2011 Pearson Addison-Wesley. All rights reserved. 18

Table 8.2b © 2011 Pearson Addison-Wesley. All rights reserved. 19

Table 8.2c © 2011 Pearson Addison-Wesley. All rights reserved. 20

Table 8.2d © 2011 Pearson Addison-Wesley. All rights reserved. 21

Table 8.3a © 2011 Pearson Addison-Wesley. All rights reserved. 22

Table 8.3b © 2011 Pearson Addison-Wesley. All rights reserved. 23

Key Terms from Chapter 8 Perfect multicollinearity Severe imperfect multicollinearity Dominant variable Auxiliary (or secondary) equation Variance inflation factor Redundant variable © 2011 Pearson Addison-Wesley. All rights reserved. 24