Heteroskedasticity.

Slides:



Advertisements
Similar presentations
Autocorrelation and Heteroskedasticity
Advertisements

Regression Analysis.
Heteroskedasticity Hill et al Chapter 11. Predicting food expenditure Are we likely to be better at predicting food expenditure at: –low incomes; –high.
Applied Econometrics Second edition
Multiple Regression Analysis
The Simple Regression Model
The Multiple Regression Model.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Homoscedasticity equal error variance. One of the assumption of OLS regression is that error terms have a constant variance across all value so f independent.
CHAPTER 8 MULTIPLE REGRESSION ANALYSIS: THE PROBLEM OF INFERENCE
Heteroskedasticity Prepared by Vera Tabakova, East Carolina University.
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
HETEROSKEDASTICITY Chapter 8.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Topic 3: Regression.
Review.
Economics Prof. Buckles
Inference for regression - Simple linear regression
Returning to Consumption
What does it mean? The variance of the error term is not constant
12.1 Heteroskedasticity: Remedies Normality Assumption.
Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
Heteroskedasticity Adapted from Vera Tabakova’s notes ECON 4551 Econometrics II Memorial University of Newfoundland.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Principles of Econometrics, 4t h EditionPage 1 Chapter 8: Heteroskedasticity Chapter 8 Heteroskedasticity Walter R. Paczkowski Rutgers University.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Chap 8 Heteroskedasticity
11.1 Heteroskedasticity: Nature and Detection Aims and Learning Objectives By the end of this session students should be able to: Explain the nature.
Multiple Regression Analysis: Inference
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Heteroscedasticity Chapter 8
Multiple Regression Analysis: Inference
Ch5 Relaxing the Assumptions of the Classical Model
Multiple Regression Analysis with Qualitative Information
Linear Regression with One Regression
Ch. 2: The Simple Regression Model
26134 Business Statistics Week 5 Tutorial
Multiple Regression Analysis: Estimation
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
REGRESSION DIAGNOSTIC II: HETEROSCEDASTICITY
Testing Multiple Linear Restrictions: The F-test
More on Specification and Data Issues
Multiple Regression Analysis with Qualitative Information
Fundamentals of regression analysis 2
More on Specification and Data Issues
Ch. 2: The Simple Regression Model
I271B Quantitative Methods
Multiple Regression Analysis with Qualitative Information
Autocorrelation.
Serial Correlation and Heteroskedasticity in Time Series Regressions
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Serial Correlation and Heteroscedasticity in
Tutorial 1: Misspecification
Chapter 7: The Normality Assumption and Inference with OLS
Seminar in Economics Econ. 470
Product moment correlation
The Simple Regression Model
Multiple Regression Analysis with Qualitative Information
Linear Panel Data Models
Multiple Regression Analysis: OLS Asymptotics
Multiple Regression Analysis: OLS Asymptotics
The Multiple Regression Model
Autocorrelation.
More on Specification and Data Issues
Tutorial 2: Autocorrelation
Financial Econometrics Fin. 505
Serial Correlation and Heteroscedasticity in
Multiple Regression Analysis
Presentation transcript:

Heteroskedasticity

Heteroskedasticity Consequences of heteroskedasticity for OLS OLS still unbiased and consistent under heteroskedastictiy! Also, interpretation of R-squared is not changed Heteroskedasticity invalidates variance formulas for OLS estimators The usual F tests and t tests are not valid under heteroskedasticity Under heteroskedasticity, OLS is no longer the best linear unbiased estimator (BLUE); there may be more efficient linear estimators Unconditional error variance is unaffected by heteroskedasticity (which refers to the conditional error variance)

Heteroskedasticity Heteroskedasticity-robust inference after OLS estimation Formulas for OLS standard errors and related statistics have been developed that are robust to heteroskedasticity of unknown form All formulas are only valid in large samples Formula for heteroskedasticity-robust OLS standard error Using these formulas, the usual t test is valid asymptotically The usual F statistic does not work under heteroskedasticity, but heteroskedasticity robust versions are available in most software Also called White/Huber/Eicker standard errors. They involve the squared residuals from the regression and from a regression of xj on all other explanatory variables.

Heteroskedasticity Example: Hourly wage equation Heteroskedasticity robust standard errors may be larger or smaller than their nonrobust counterparts. The differences are often small in practice. F statistics are also often not too different. If there is strong heteroskedasticity, differences may be larger. To be on the safe side, it is advisable to always compute robust standard errors.

Heteroskedasticity Testing for heteroskedasticity It may still be interesting whether there is heteroskedasticity because then OLS may not be the most efficient linear estimator anymore Breusch-Pagan test for heteroskedasticity Under MLR.4 The mean of u2 must not vary with x1, x2, …, xk

Heteroskedasticity Breusch-Pagan test for heteroskedasticity (cont.) Regress squared residuals on all expla-natory variables and test whether this regression has explanatory power. A large test statistic (= a high R-squared) is evidence against the null hypothesis. Alternative test statistic (= Lagrange multiplier statistic, LM). Again, high values of the test statistic (= high R-squared) lead to rejection of the null hypothesis that the expected value of u2 is unrelated to the explanatory variables.

Heteroskedasticity Example: Heteroskedasticity in housing price equations Heteroskedasticity In the logarithmic specification, homoskedasticity cannot be rejected

Heteroskedasticity The White test for heteroskedasticity Disadvantage of this form of the White test Including all squares and interactions leads to a large number of estimated parameters (e.g. k=6 leads to 27 parameters to be estimated) Regress squared residuals on all expla-natory variables, their squares, and in-teractions (here: example for k=3) The White test detects more general deviations from heteroskedasticity than the Breusch-Pagan test

Heteroskedasticity Alternative form of the White test Example: Heteroskedasticity in (log) housing price equations This regression indirectly tests the dependence of the squared residuals on the explanatory variables, their squares, and interactions, because the predicted value of y and its square implicitly contain all of these terms.

Heteroskedasticity Weighted least squares estimation Heteroskedasticity is known up to a multiplicative constant The functional form of the heteroskedasticity is known Transformed model

Heteroskedasticity Example: Savings and income The transformed model is homoskedastic If the other Gauss-Markov assumptions hold as well, OLS applied to the transformed model is the best linear unbiased estimator Note that this regression model has no intercept

Heteroskedasticity OLS in the transformed model is weighted least squares (WLS) Why is WLS more efficient than OLS in the original model? Observations with a large variance are less informative than observations with small variance and therefore should get less weight WLS is a special case of generalized least squares (GLS) Observations with a large variance get a smaller weight in the optimization problem

Heteroskedasticity Example: Financial wealth equation Net financial wealth Assumed form of heteroskedasticity WLS estimates have considerably smaller standard errors (which is line with the expectation that they are more efficient). Participation in 401K pension plan

Heteroskedasticity Important special case of heteroskedasticity If the observations are reported as averages at the city/county/state/-country/firm level, they should be weighted by the size of the unit Average contribution to pension plan in firm i Average earnings and age in firm i Percentage firm contributes to plan Heteroskedastic error term Error variance if errors are homoskedastic at the individual-level If errors are homoskedastic at the individual-level, WLS with weights equal to firm size mi should be used. If the assumption of homoskedasticity at the individual-level is not exactly right, one can calculate robust standard errors after WLS (i.e. for the transformed model).

Heteroskedasticity Unknown heteroskedasticity function (feasible GLS) Assumed general form of heteroskedasticity; exp-function is used to ensure positivity Multiplicative error (assumption: independent of the explanatory variables) Use inverse values of the estimated heteroskedasticity funtion as weights in WLS Feasible GLS is consistent and asymptotically more efficient than OLS.

Heteroskedasticity Example: Demand for cigarettes Estimation by OLS Smoking restrictions in restaurants Cigarettes smoked per day Logged income and cigarette price Reject homo-skedasticity

Heteroskedasticity Estimation by FGLS Discussion The income elasticity is now statistically significant; other coefficients are also more precisely estimated (without changing qualit. results) Now statistically significant

Heteroskedasticity What if the assumed heteroskedasticity function is wrong? If the heteroskedasticity function is misspecified, WLS is still consistent under MLR.1 – MLR.4, but robust standard errors should be computed WLS is consistent under MLR.4 but not necessarily under MLR.4‘ If OLS and WLS produce very different estimates, this typically indi- cates that some other assumptions (e.g. MLR.4) are wrong If there is strong heteroskedasticity, it is still often better to use a wrong form of heteroskedasticity in order to increase efficiency

Heteroskedasticity WLS in the linear probability model Discussion Infeasible if LPM predictions are below zero or greater than one If such cases are rare, they may be adjusted to values such as .01/.99 Otherwise, it is probably better to use OLS with robust standard errors In the LPM, the exact form of heteroskedasticity is known Use inverse values as weights in WLS