1 Heteroskedasticity. 2 The Nature of Heteroskedasticity  Heteroskedasticity is a systematic pattern in the errors where the variances of the errors.

Slides:



Advertisements
Similar presentations
Autocorrelation and Heteroskedasticity
Advertisements

Heteroskedasticity Hill et al Chapter 11. Predicting food expenditure Are we likely to be better at predicting food expenditure at: –low incomes; –high.
Assumptions underlying regression analysis
Applied Econometrics Second edition
Heteroskedasticity Lecture 17 Lecture 17.
Properties of Least Squares Regression Coefficients
Econometric Modeling Through EViews and EXCEL
The Multiple Regression Model.
Homoscedasticity equal error variance. One of the assumption of OLS regression is that error terms have a constant variance across all value so f independent.
Heteroskedasticity Prepared by Vera Tabakova, East Carolina University.
Heteroskedasticity The Problem:
Objectives (BPS chapter 24)
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Multiple regression analysis
Chapter 10 Simple Regression.
2.5 Variances of the OLS Estimators
CHAPTER 3 ECONOMETRICS x x x x x Chapter 2: Estimating the parameters of a linear regression model. Y i = b 1 + b 2 X i + e i Using OLS Chapter 3: Testing.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Chapter 5 Heteroskedasticity. What is in this Chapter? How do we detect this problem What are the consequences of this problem? What are the solutions?
Statistical Background
1 MF-852 Financial Econometrics Lecture 6 Linear Regression I Roy J. Epstein Fall 2003.
Econ 140 Lecture 191 Heteroskedasticity Lecture 19.
Review.
Economics Prof. Buckles
Statistics 350 Lecture 17. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Chapter 12 Section 1 Inference for Linear Regression.
ECON 7710, Heteroskedasticity What is heteroskedasticity? What are the consequences? How is heteroskedasticity identified? How is heteroskedasticity.
Inference for regression - Simple linear regression
Hypothesis Testing in Linear Regression Analysis
Returning to Consumption
Quantitative Methods Heteroskedasticity.
What does it mean? The variance of the error term is not constant
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
12.1 Heteroskedasticity: Remedies Normality Assumption.
Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1Spring 02 Problems in Regression Analysis Heteroscedasticity Violation of the constancy of the variance of the errors. Cross-sectional data Serial Correlation.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Heteroskedasticity Adapted from Vera Tabakova’s notes ECON 4551 Econometrics II Memorial University of Newfoundland.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Heteroskedasticity ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Chapter 2 Statistical Background. 2.3 Random Variables and Probability Distributions A variable X is said to be a random variable (rv) if for every real.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
EC 532 Advanced Econometrics Lecture 1 : Heteroscedasticity Prof. Burak Saltoglu.
Principles of Econometrics, 4t h EditionPage 1 Chapter 8: Heteroskedasticity Chapter 8 Heteroskedasticity Walter R. Paczkowski Rutgers University.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)
Chap 8 Heteroskedasticity
Stat 112 Notes 14 Assessing the assumptions of the multiple regression model and remedies when assumptions are not met (Chapter 6).
11.1 Heteroskedasticity: Nature and Detection Aims and Learning Objectives By the end of this session students should be able to: Explain the nature.
Heteroskedasticity ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Lecturer: Ing. Martina Hanová, PhD..  How do we evaluate a model?  How do we know if the model we are using is good?  assumptions relate to the (population)
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Heteroscedasticity Chapter 8
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
REGRESSION DIAGNOSTIC II: HETEROSCEDASTICITY
Fundamentals of regression analysis 2
CHAPTER 29: Multiple Regression*
Autocorrelation.
Heteroskedasticity.
Autocorrelation Dr. A. PHILIP AROKIADOSS Chapter 5 Assistant Professor
Autocorrelation.
Heteroskedasticity.
Financial Econometrics Fin. 505
Regression Models - Introduction
Presentation transcript:

1 Heteroskedasticity

2 The Nature of Heteroskedasticity  Heteroskedasticity is a systematic pattern in the errors where the variances of the errors are not constant.  Ordinary least squares assumes that all observations are equally reliable (constant variance).  For efficiency (accurate estimation / prediction) re-weight observations to ensure equal error variance.

3 y t =  1 +  2 x t + ε t Regression Model E( ε t ) = 0 var( ε t ) =  2 zero mean: homoskedasticity: nonautocorrelation: heteroskedasticity: var( ε t ) =  t 2 cov(ε t, ε s ) =  t = s

4 Homoskedastic pattern of errors xtxt ytyt Income Consumption

5.. xtxt x1x1 x2x2 ytyt f(y t ) The Homoskedastic Case.. x3x3 x4x4 Income Consumption

6 Heteroskedastic pattern of errors xtxt ytyt income consumption

7. x t x1x1 x2x2 ytyt f(y t ) Consumption x3x3.. The Heteroskedastic Case Income Rich people Poor people

8 Properties of Least Squares 1. Least squares still linear and unbiased. 2.Least squares NOT efficient. 3.Hence, it is no longer B.L.U.E. 4. Usual formulas give incorrect standard errors for least squares. 5. Confidence intervals and hypothesis tests based on usual standard errors are wrong.

9 y t =  1 +  2 x t + ε Heteroskedasticity: E(ε t ) = 0, var(ε t ) =  t 2, cov(ε t, ε s ) =  t  s  (Linear) Where  (Unbiased)

10 y t =  1 +  2 x t + ε t heteroskedasticity: var(ε t ) =  t  2 Incorrect formula for least squares variance: var(b 2 ) = 22  x t  x   Correct formula for least squares variance: var(b 2 ) =   t 2  x t  x    x t  x    

11 Halbert White Standard Errors White estimator of the least squares variance: est.var(b 2 ) =  ε t 2  x t  x    x t  x     ^ 1.In large samples, White standard error (square root of estimated variance) is a consistent measure. 2. Because the squared residuals are used to approximate the variances, White's estimator is strictly appropriate only in large samples. : the squares of the least squares residuals εt2εt2 ^

12 Two Types of Heteroskedasticity 1.Proportional Heteroskedasticity. (continuous function (of x t, for example) ) For example, Income is less important as an explanatory variable for food expenditure of high- income families. It is harder to guess their food expenditure. 2.Partitioned Heteroskedasticity. (discrete categories/groups) For instance, exchange rates are more volatile after Asian Financial Crisis.

13 Proportional Heteroskedasticity y t =  1 +  2 x t + ε t where  t  2 =   2 x t The variance is assumed to be proportional to the value of x t var( ε t ) =  t  2 E(ε t ) = 0 cov(ε t, ε s ) = 0 t = s

14  t  2 =   2 x t y t =  1 +  2 x t + ε t std.dev. proportional to x t variance: standard deviation:  t  =   x t y t 1 x t ε t =  1 +  2 + x t x t To correct for heteroskedasticity divide the model by x t var(ε t ) =  t  2 It is important to recognize that  1 and  2 are the same in both the transformed model and the untransformed model

15 y t 1 x t ε t =  1 +  2 + x t x t y t =  1 x t1 +  2 x t2 + ε t * ** * var( ε t )= var( ) = var( ε t )=   2 x t * εtεt xtxt 1 xtxt 1 x t var( ε t ) =   2 * ε t is heteroskedastic, but ε t is homoskedastic. *

16 1. Decide which variable is proportional to the heteroskedasticity ( x t in previous example). 2. Divide all terms in the original model by the square root of that variable (divide by x t ). 3. Run least squares on the transformed model which has new y t, x t1 and x t2 variables but no intercept. Generalized Least Squares These steps describe weighted least squares: ***

17  The errors are weighted by the reciprocal of x t.  When x t is small, the data contain more information about the regression function and the observations are weighted heavily.  When x t is large, the data contain less information and the observations are weighted lightly.  In this way we take advantage of the heteroskedasticity to improve parameter estimation (efficiency).

18 Partitioned Heteroskedasticity y t =  1 +  2 x t + ε t var( ε t ) =  1 2 var( ε t ) =  2 2 Error variance of field corn : Error variance of sweet corn : y t = bushels per acre of corn x t = gallons of water per acre (rain or other) t = 1,...,100 t = 1,...,80 t = 81,...,100

19 y t =  1 +  2 x t + ε t var( ε t ) =  1 2 Field corn : y t =  1 +  2 x t + ε t var( ε t ) =  2 2 Sweet corn : y t 1 x t ε t =  1 +  2 +  1  1 y t 1 x t ε t =  1 +  2 +  2  2 Re-weighting Each Group’s Observations t = 1,...,80 t = 81,...,100

20 y t =  1 x t1 +  2 x t2 + ε t * ** * var( ε t ) = var( ) = var( ε t ) = * εtεt 1 1 var( ε t ) = 1 * ε t is heteroskedastic, but ε t is homoskedastic. * y t 1 x t ε t =  1 +  2 +  i  i  i  i  i 2 t = 1,...,100

21 Apply Generalized Least Squares Run least squares separately on data for each group.  1 2 (MSE 1 ) provides estimator of  1 2 using the 80 observations on field corn. ^  2 2 (MSE 2 ) provides estimator of  2 2 using the 20 observations on sweet corn. ^

22 1. Residual Plots provide information on the exact nature of heteroskedasticity (partitioned or proportional) to aid in correcting for it. 2. Goldfeld-Quandt Test checks for presence of heteroskedasticity. Detecting Heteroskedasticity Determine existence and nature of heteroskedasticity:

23 Residual Plots etet 0 xtxt Plot residuals against one variable at a time after sorting the data by that variable to try to find a heteroskedastic pattern in the data.

24 Goldfeld-Quandt Test The Goldfeld-Quandt test can be used to detect heteroskedasticity in either the proportional case or for comparing two groups in the discrete case. For proportional heteroskedasticity, it is first necessary to determine which variable, such as x t, is proportional to the error variance. Then sort the data from the largest to smallest values of that variable.

25 H o :  1  2 =  2  2 H 1 :  1  2 >  2  2 GQ = ~ F [T 1  K 1, T 2  K 2 ] 1212 2222 ^ ^ In the proportional case, (drop the middle r observations where r  T/6,) run separate least squares regressions on the first T 1 observations and the last T 2 observations. We assume that  1 2 >  2 2. (If not, then reverse the subscripts.) The Small values of GQ support H o while large values support H 1. Goldfeld-Quandt Test Statistic