1 Javier Aparicio División de Estudios Políticos, CIDE Primavera 2009 Regresión.

Slides:



Advertisements
Similar presentations
Applied Econometrics Second edition
Advertisements

Heteroskedasticity Lecture 17 Lecture 17.
Multiple Regression Analysis
The Simple Regression Model
1 Javier Aparicio División de Estudios Políticos, CIDE Otoño Regresión.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
Lecture 8 (Ch14) Advanced Panel Data Method
Heteroskedasticity Prepared by Vera Tabakova, East Carolina University.
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
The Simple Linear Regression Model: Specification and Estimation
HETEROSKEDASTICITY Chapter 8.
Cross section and panel method
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
2.5 Variances of the OLS Estimators
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
Prof. Dr. Rainer Stachuletz
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Chapter 5 Heteroskedasticity. What is in this Chapter? How do we detect this problem What are the consequences of this problem? What are the solutions?
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Multiple Regression Analysis
Multiple Regression Analysis
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 3. Asymptotic Properties.
Econ 140 Lecture 191 Heteroskedasticity Lecture 19.
Topic 3: Regression.
The Simple Regression Model
Review.
EC Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Economics Prof. Buckles
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
ECON 7710, Heteroskedasticity What is heteroskedasticity? What are the consequences? How is heteroskedasticity identified? How is heteroskedasticity.
Class 4 Ordinary Least Squares SKEMA Ph.D programme Lionel Nesta Observatoire Français des Conjonctures Economiques
Returning to Consumption
Quantitative Methods Heteroskedasticity.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
12.1 Heteroskedasticity: Remedies Normality Assumption.
Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
1 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved. CH5 Multiple Regression Analysis: OLS Asymptotic 
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Heteroskedasticity Adapted from Vera Tabakova’s notes ECON 4551 Econometrics II Memorial University of Newfoundland.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Heteroskedasticity ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Principles of Econometrics, 4t h EditionPage 1 Chapter 8: Heteroskedasticity Chapter 8 Heteroskedasticity Walter R. Paczkowski Rutgers University.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)
Chap 8 Heteroskedasticity
Quantitative Methods. Bivariate Regression (OLS) We’ll start with OLS regression. Stands for  Ordinary Least Squares Regression. Relatively basic multivariate.
Heteroskedasticity ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Estimation Econometría. ADE.. Estimation We assume we have a sample of size T of: – The dependent variable (y) – The explanatory variables (x 1,x 2, x.
High Speed Heteroskedasticity Review. 2 Review: Heteroskedasticity Heteroskedasticity leads to two problems: –OLS computes standard errors on slopes incorrectly.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Heteroscedasticity Chapter 8
Multiple Regression Analysis: Estimation
The Simple Regression Model
Multiple Regression Analysis
I271B Quantitative Methods
Simple Linear Regression
Tutorial 1: Misspecification
Heteroskedasticity.
Multiple Regression Analysis
Presentation transcript:

1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión Lineal Múltiple y i =  0 +  1 x 1i +  2 x 2i +...  k x ki + u i Ch 8. Heteroskedasticity

2 What is Heteroskedasticity Recall the assumption of homoskedasticity implied that conditional on the explanatory variables, the variance of the unobserved error, u, was constant If this is not true, that is if the variance of u is different for different values of the x’s, then the errors are heteroskedastic Example: estimating returns to education and ability is unobservable, and think the variance in ability differs by educational attainment

3. xx1x1 x2x2 y f(y|x) Example of Heteroskedasticity x3x3.. E(y|x) =  0 +  1 x

4 Why Worry About Heteroskedasticity? OLS is still unbiased and consistent, even if we do not assume homoskedasticity The standard errors of the estimates are biased if we have heteroskedasticity If the standard errors are biased, we can not use the usual t statistics or F statistics or LM statistics for drawing inferences

5 Variance with Heteroskedasticity

6

7 Robust Standard Errors Now that we have a consistent estimate of the variance, the square root can be used as a standard error for inference Typically call these robust standard errors Sometimes the estimated variance is corrected for degrees of freedom by multiplying by n/(n – k – 1) As n → ∞ it’s all the same, though

8 Robust Standard Errors (cont) Important to remember that these robust standard errors only have asymptotic justification – with small sample sizes t statistics formed with robust standard errors will not have a distribution close to the t, and inferences will not be correct In Stata, robust standard errors are easily obtained using the robust option of reg

9 A Robust LM Statistic Run OLS on the restricted model and save the residuals ŭ Regress each of the excluded variables on all of the included variables (q different regressions) and save each set of residuals ř 1, ř 2, …, ř q Regress a variable defined to be = 1 on ř 1 ŭ, ř 2 ŭ, …, ř q ŭ, with no intercept The LM statistic is n – SSR 1, where SSR 1 is the sum of squared residuals from this final regression

10 Testing for Heteroskedasticity Essentially want to test H 0 : Var(u|x 1, x 2,…, x k ) =  2, which is equivalent to H 0 : E(u 2 |x 1, x 2,…, x k ) = E(u 2 ) =  2 If assume the relationship between u 2 and x j will be linear, can test as a linear restriction So, for u 2 =  0 +  1 x 1 +…+  k x k + v this means testing H 0 :  1 =  2 = … =  k = 0

11 The Breusch-Pagan Test Don’t observe the error, but can estimate it with the residuals from the OLS regression After regressing the residuals squared on all of the x’s, can use the R 2 to form an F or LM test The F statistic is just the reported F statistic for overall significance of the regression, F = [R 2 /k]/[(1 – R 2 )/(n – k – 1)], which is distributed F k, n – k - 1 The LM statistic is LM = nR 2, which is distributed  2 k Use bpagan package in stata (findit bpagan)

12 The White Test The Breusch-Pagan test will detect any linear forms of heteroskedasticity The White test allows for nonlinearities by using squares and crossproducts of all the x’s Still just using an F or LM to test whether all the x j, x j 2, and x j x h are jointly significant This can get to be unwieldy pretty quickly Use whitetst package in stata (findit whitetst)

13 Alternate form of the White test Consider that the fitted values from OLS, ŷ, are a function of all the x’s Thus, ŷ 2 will be a function of the squares and crossproducts and ŷ and ŷ 2 can proxy for all of the x j, x j 2, and x j x h, so Regress the residuals squared on ŷ and ŷ 2 and use the R 2 to form an F or LM statistic Note only testing for 2 restrictions now

14 Weighted Least Squares While it’s always possible to estimate robust standard errors for OLS estimates, if we know something about the specific form of the heteroskedasticity, we can obtain more efficient estimates than OLS The basic idea is going to be to transform the model into one that has homoskedastic errors – called weighted least squares See rreg command in stata

15 Case of form being known up to a multiplicative constant Suppose the heteroskedasticity can be modeled as Var(u|x) =  2 h(x), where the trick is to figure out what h(x) ≡ h i looks like E(u i /√h i |x) = 0, because h i is only a function of x, and Var(u i /√h i |x) =  2, because we know Var(u|x) =  2 h i So, if we divided our whole equation by √h i we would have a model where the error is homoskedastic

16 Generalized Least Squares Estimating the transformed equation by OLS is an example of generalized least squares (GLS) GLS will be BLUE in this case GLS is a weighted least squares (WLS) procedure where each squared residual is weighted by the inverse of Var(u i |x i )

17 Weighted Least Squares While it is intuitive to see why performing OLS on a transformed equation is appropriate, it can be tedious to do the transformation Weighted least squares is a way of getting the same thing, without the transformation Idea is to minimize the weighted sum of squares (weighted by 1/h i )

18 More on WLS WLS is great if we know what Var(u i |x i ) looks like In most cases, won’t know form of heteroskedasticity Example where do is if data is aggregated, but model is individual level Want to weight each aggregate observation by the inverse of the number of individuals

19 Feasible GLS More typical is the case where you don’t know the form of the heteroskedasticity In this case, you need to estimate h(x i ) Typically, we start with the assumption of a fairly flexible model, such as Var(u|x) =  2 exp(  0 +  1 x 1 + …+  k x k ) Since we don’t know the , must estimate

20 Feasible GLS (continued) Our assumption implies that u 2 =  2 exp(  0 +  1 x 1 + …+  k x k )v Where E(v|x) = 1, then if E(v) = 1 ln(u 2 ) =    +  1 x 1 + …+  k x k + e Where E(e) = 1 and e is independent of x Now, we know that û is an estimate of u, so we can estimate this by OLS

21 Feasible GLS (continued) Now, an estimate of h is obtained as ĥ = exp(ĝ), and the inverse of this is our weight So, what did we do? Run the original OLS model, save the residuals, û, square them and take the log Regress ln(û 2 ) on all of the independent variables and get the fitted values, ĝ Do WLS using 1/exp(ĝ) as the weight

22 WLS Wrapup When doing F tests with WLS, form the weights from the unrestricted model and use those weights to do WLS on the restricted model as well as the unrestricted model Remember we are using WLS just for efficiency – OLS is still unbiased & consistent Estimates will still be different due to sampling error, but if they are very different then it’s likely that some other Gauss-Markov assumption is false