8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)

Slides:



Advertisements
Similar presentations
Autocorrelation and Heteroskedasticity
Advertisements

Heteroskedasticity Hill et al Chapter 11. Predicting food expenditure Are we likely to be better at predicting food expenditure at: –low incomes; –high.
Applied Econometrics Second edition
Properties of Least Squares Regression Coefficients
Econometric Modeling Through EViews and EXCEL
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Fourteen.
Homoscedasticity equal error variance. One of the assumption of OLS regression is that error terms have a constant variance across all value so f independent.
Chapter 11 Autocorrelation.
Heteroskedasticity Prepared by Vera Tabakova, East Carolina University.
Introduction and Overview
Multicollinearity Multicollinearity - violation of the assumption that no independent variable is a perfect linear function of one or more other independent.
Heteroskedasticity The Problem:
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
Class 16: Thursday, Nov. 4 Note: I will you some info on the final project this weekend and will discuss in class on Tuesday.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Multiple Linear Regression Model
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Chapter 5 Heteroskedasticity. What is in this Chapter? How do we detect this problem What are the consequences of this problem? What are the solutions?
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
Econ 140 Lecture 181 Multiple Regression Applications III Lecture 18.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Lecture 24: Thurs., April 8th
Statistics 350 Lecture 10. Today Last Day: Start Chapter 3 Today: Section 3.8 Homework #3: Chapter 2 Problems (page 89-99): 13, 16,55, 56 Due: February.
Topic 3: Regression.
Review.
Ch. 14: The Multiple Regression Model building
Economics Prof. Buckles
Simple Linear Regression Analysis
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Ordinary Least Squares
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
ECON 7710, Heteroskedasticity What is heteroskedasticity? What are the consequences? How is heteroskedasticity identified? How is heteroskedasticity.
Introduction to Linear Regression and Correlation Analysis
Hypothesis Testing in Linear Regression Analysis
Returning to Consumption
Quantitative Methods Heteroskedasticity.
What does it mean? The variance of the error term is not constant
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Soc 3306a Multiple Regression Testing a Model and Interpreting Coefficients.
1 MF-852 Financial Econometrics Lecture 10 Serial Correlation and Heteroscedasticity Roy J. Epstein Fall 2003.
Pure Serial Correlation
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
12.1 Heteroskedasticity: Remedies Normality Assumption.
Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Principles of Econometrics, 4t h EditionPage 1 Chapter 8: Heteroskedasticity Chapter 8 Heteroskedasticity Walter R. Paczkowski Rutgers University.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Chap 8 Heteroskedasticity
10-1 MGMG 522 : Session #10 Simultaneous Equations (Ch. 14 & the Appendix 14.6)
5-1 MGMG 522 : Session #5 Multicollinearity (Ch. 8)
7-1 MGMG 522 : Session #7 Serial Correlation (Ch. 9)
Quantitative Methods. Bivariate Regression (OLS) We’ll start with OLS regression. Stands for  Ordinary Least Squares Regression. Relatively basic multivariate.
11.1 Heteroskedasticity: Nature and Detection Aims and Learning Objectives By the end of this session students should be able to: Explain the nature.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
Heteroscedasticity Heteroscedasticity is present if the variance of the error term is not a constant. This is most commonly a problem when dealing with.
Multiple Regression Analysis: Estimation
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
REGRESSION DIAGNOSTIC II: HETEROSCEDASTICITY
Fundamentals of regression analysis 2
STOCHASTIC REGRESSORS AND THE METHOD OF INSTRUMENTAL VARIABLES
Pure Serial Correlation
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Undergraduated Econometrics
Heteroskedasticity.
Financial Econometrics Fin. 505
Presentation transcript:

8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)

8-2 Heteroskedasticity  Heteroskedasticity is a violation of the classical assumption #5 (observations of the error term are drawn from a distribution that has a constant variance). (See Figure 10.1 on p. 348)  Heteroskedasticity can be categorized into 2 kinds: –Pure heteroskedasticity (Heteroskedasticity that exists in a correctly specified regression model). –Impure heteroskedasticity (Heteroskedasticity that is caused by specification errors: omitted variables or incorrect functional form).  Heteroskedasticity mostly happens in a cross- sectional data set. This, however, doesn’t mean that time-series data cannot have heteroskedasticity.

8-3 Pure Heteroskedasticity  Classical assumption #5 implies that the observations of the error term are drawn from a distribution with a constant variance, or VAR(ε i ) =  2.  In other words, the error term must be homoskedastic.  For a heteroskedastic error term, VAR(ε i ) =  i 2.  Heteroskedasticity can take on many forms but we will limit our discussion to one form of heteroskedasticity only.

8-4 Heteroskedasticity of the Form: VAR(ε i ) =  2 Z i 2  Z is called a proportionality factor because the variance of the error term changes proportionally to the Z factor (or variable). (See Figure 10.2 on p. 350 and Figure 10.3 on p. 351)  The variance of the error term increases as Z increases.  Z may or may not be one of the explanatory variables in the regression model.  Heteroskedasticity can occur in at least 2 situations: 1. When there is significant change in the dependent variable of a time-series model. 2. When there are different amounts of measurement errors in the sample of different periods or different sub-samples.

8-5 Impure Heteroskedasticity  Caused by specification errors, especially the omitted variables.  Incorrect functional form is less likely to cause heteroskedasticity.  Specification errors should be corrected first by way of investigating the independent variables and/or functional form.

8-6 How can omitted variables cause heteroskedasticity?  Correct model: Y =  0 + 1 X 1 + 2 X 2 +ε  If X 2 is omitted: Y =  0 + 1 X 1 +ε *  Where, ε * =  2 X 2 +ε  If ε is small compared to  2 X 2, and X 2 is heteroskedastic, the ε * will also be heteroskedastic.  Both the bias and impure heteroskedasticity will disappear once the model gets corrected.

8-7 Consequences of Heteroskedasticity 1. Pure heteroskedasticity does not cause bias in the coefficient estimates. 2. Pure heteroskedasticity increases variances of the coefficient estimates. 3. Pure heteroskedasticity causes OLS to underestimate the standard errors of the coefficients. (Hence, pure heteroskedasticity overestimates the t- values.)

8-8 Detection of Heteroskedasticity  Because heteroskedasticity can take on many forms, therefore, there is no specific test to test for heteroskedasticity.  However, we will discuss two tests as a tool for detecting heteroskedasticity of a certain form.  The two tests are (1) Park Test, and (2) White Test.

8-9 Detection of Heteroskedasticity: Caveats  Before you go on and conduct one or both of these tests, ask yourself these 3 questions: 1. Is your model already correctly specified? If the answer is “No”, correct your model first. 2. Does the topic you are studying have a known heteroskedasticity problem as pointed out by other researchers? 3. When you plot a graph of the residuals against the Z factor, does it indicate a potential heteroskedasticity problem?

8-10 The Park Test  The Park test tests for heteroskedasticity of the form, VAR(ε i ) =  2 Z i 2.  One difficulty with the Park test is the specification of the Z factor.  The Z factor is usually one of the explanatory variables, but not always. The Z factor could very well be other variable not included in the regression model.

Steps of the Park Test 1. Run the OLS and get the residuals: e i = Y i –b 0 –b 1 X 1i –b 2 X 2i 2. Regress the ln(e) 2 on the ln(Z): ln(e i 2 ) = b 0 +b 1 ln(Z i ) 3. Check to see if b 1 is significant or not. –If b 1 is significant (meaning b 1 ≠ 0), this implies heteroskedasticity. –If b 1 is not significant (meaning b 1 = 0), this implies that heteroskedasticity is unlikely. (However, it may still exist in other form or exist with other Z factor.)

8-12 The White Test  In stead of having to specify the Z factor and heteroskedasticity of the form VAR(ε i ) =  2 Z i 2, the White test takes a more general approach.

Steps in the White Test 1. Run the OLS and get the residuals: e i = Y i –b 0 –b 1 X 1i –b 2 X 2i –b 3 X 3i 2. Regress e 2 on each X, square of each X, cross product of the Xs. For a 3-independent variable model, run OLS of 3. Test the overall significance of the model above with  2 test. The test statistic is nR 2. –If nR 2 > critical  2 value, this implies heteroskedasticity. –If nR 2 < critical  2 value, this implies that heteroskedasticity is unlikely.  n = number of observations, R 2 = the value of the unadjusted R 2, the DF for the critical  2 value is the number of the independent variables in the model above.

8-14 Correcting Heteroskedasticity  If heteroskedasticity still is present in your correctly specified regression model (pure heteroskedasticity), consider one of these three remedies. 1. Use Weighted Least Squares 2. Use Heteroskedasticity-Corrected Standard Errors 3. Redefine the variables (See an example on p )

8-15 Weighted Least Squares  Take an original model with a heteroskedastic error term,  Suppose the variance is of the form,  Eq.1 can be shown to be equal to,  Divide Eq.2 with Z i, obtaining See 10-6 on p. 598 for proof.

8-16 Weighted Least Squares (Con..)  If Z happens to be one of the independent variables, use OLS with Eq.3 specification to obtain the coefficient estimates.  If Z is not one of the independent variables, we must add an intercept term because Eq.3 has no intercept term. We will run OLS using Eq.4 specification.  The interpretation of the slope coefficients becomes a bit tricky.  There are two problems with WLS: (1) what is Z?, and (2) how Z relates to VAR(ε)?

8-17 White’s Heteroskedasticity-Corrected Standard Errors (a.k.a. Robust Standard Errors)  The idea of this remedy goes like this.  Since, there is no bias in the coefficient estimates.  But, the standard errors of the coefficient estimates are larger with heteroskedasticity than without it.  Therefore, why not fix the standard errors and leave the coefficient estimates alone?  This method is referred to as HCCM (heteroskedasticity-consistent covariance matrix).  In EViews, you will choose “LS” and click on “Options”, then select “Heteroskedasticity- Consistent Coefficient Covariance” and select “White”.  The standard errors of the coefficient estimates can be bigger or smaller than those from the OLS.  This method works best in a large sample data.

8-18 Redefine the Variables 1. For example, take log of some variables. 2. Or, normalize some variables.