Autocorrelation.

Slides:



Advertisements
Similar presentations
Regression Analysis.
Advertisements

Heteroskedasticity Hill et al Chapter 11. Predicting food expenditure Are we likely to be better at predicting food expenditure at: –low incomes; –high.
Properties of Least Squares Regression Coefficients
Autocorrelation Lecture 20 Lecture 20.
Econometric Modeling Through EViews and EXCEL
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Managerial Economics in a Global Economy
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Notes. What is a simple linear relation? When one variable is associated with another variable in such a way that two numbers completely.
Objectives (BPS chapter 24)
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 13 Additional Topics in Regression Analysis
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Additional Topics in Regression Analysis
Chapter 11 Multiple Regression.
Topic 3: Regression.
Autocorrelation Lecture 18 Lecture 18.
Simple Linear Regression Analysis
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Inference for regression - Simple linear regression
Regression Method.
Returning to Consumption
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Serial Correlation and the Housing price function Aka “Autocorrelation”
1 MF-852 Financial Econometrics Lecture 10 Serial Correlation and Heteroscedasticity Roy J. Epstein Fall 2003.
Pure Serial Correlation
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Principles of Econometrics, 4t h EditionPage 1 Chapter 8: Heteroskedasticity Chapter 8 Heteroskedasticity Walter R. Paczkowski Rutgers University.
Correlation & Regression Analysis
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
1 Heteroskedasticity. 2 The Nature of Heteroskedasticity  Heteroskedasticity is a systematic pattern in the errors where the variances of the errors.
11.1 Heteroskedasticity: Nature and Detection Aims and Learning Objectives By the end of this session students should be able to: Explain the nature.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Lecturer: Ing. Martina Hanová, PhD..  How do we evaluate a model?  How do we know if the model we are using is good?  assumptions relate to the (population)
Multiple Regression Analysis: Inference
Ch5 Relaxing the Assumptions of the Classical Model
Introduction For inference on the difference between the means of two populations, we need samples from both populations. The basic assumptions.
More on Inference.
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
REGRESSION DIAGNOSTIC III: AUTOCORRELATION
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Dynamic Models, Autocorrelation and Forecasting
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Chapter 4. Inference about Process Quality
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
Lecture 8 Preview: Interval Estimates and Hypothesis Testing
Econometric methods of analysis and forecasting of financial markets
Multivariate Regression
Fundamentals of regression analysis 2
Pure Serial Correlation
...Relax... 9/21/2018 ST3131, Lecture 3 ST5213 Semester II, 2000/2001
CHAPTER 29: Multiple Regression*
More on Inference.
Chapter 12 – Autocorrelation
Autocorrelation.
Serial Correlation and Heteroskedasticity in Time Series Regressions
Serial Correlation and Heteroscedasticity in
Simple Linear Regression
Heteroskedasticity.
Product moment correlation
BEC 30325: MANAGERIAL ECONOMICS
Chapter 13 Additional Topics in Regression Analysis
Autocorrelation Dr. A. PHILIP AROKIADOSS Chapter 5 Assistant Professor
Autocorrelation MS management.
BEC 30325: MANAGERIAL ECONOMICS
Serial Correlation and Heteroscedasticity in
Presentation transcript:

Autocorrelation

The Nature of Autocorrelation The randomness of the sample implies that the error terms for different observations will be uncorrelated. When we have time-series data, where the observations follow a natural ordering through time, there is always a possibility that successive errors will be correlated with each other. In any one period, the current error term contains not only the effects of current shocks but also the carryover from previous shocks. This carryover will be related to, or correlated with, the effects of the earlier shocks. When circumstances such as these lead to error terms that are correlated, we say that autocorrelation exists. The possibility of autocorrelation should always be entertained when we are dealing with time-series data.

For efficiency (accurate estimation / prediction) all systematic information needs to be incorporated into the regression model. Autocorrelation is a systematic pattern in the errors that can be either attracting (positive) or repelling (negative) autocorrelation.

crosses line not enough (attracting) et . . Postive Auto. . . . . . . . . . . . . . . . . . . . . t crosses line randomly et . . . . . . . . . . . . No Auto. . . . . . . . . . . . . . . . . t crosses line too much (repelling) . et . . . . . . . . Negative Auto. . . . . . . . t . .

Regression Model yt = 1 + 2xt + et E(et) = 0 var(et) = 2 zero mean: E(et) = 0 homoskedasticity: var(et) = 2 nonautocorrelation: cov(et, es) = t = s autocorrelation: cov(et, es) = t = s

Order of Autocorrelation yt = 1 + 2xt + et 1st Order: et = et1 + t et = 1et1 + 2et2 + t 2nd Order: 3rd Order: et = 1et1 + 2et2 + 3et3 + t We will assume First Order Autocorrelation: et = et1 + t AR(1) :

First Order Autocorrelation yt = 1 + 2xt + et et = et1 + t where 1 <  < 1 var(t) = 2 E(t) = 0 cov(t, s) = t = s These assumptions about t imply the following about et : E(et) = 0 cov(et, etk) = e2 k for k > 0 2 var(et) = e2 = corr(et, etk) = k for k > 0 12

Autocorrelation creates some Problems for Least Squares: If we have an equation whose errors exhibit autocorrelation, but we ignore it, or are simply unaware of it, what does it have on the properties of least squares estimates? 1. The least squares estimator is still linear and unbiased but it is not efficient. 2. The formulas normally used to compute the least squares standard errors are no longer correct and confidence intervals and hypothesis tests using them will be wrong.

yt = 1 + 2xt + e Autocorrelation: E(et) = 0, var(et) = 2 , cov(et, es) = ,t  s (Linear) Where (Unbiased)

xt x  yt = 1 + 2xt + et e2 cov(et, es) = ,t  s Autocorrelation: Incorrect formula for least squares variance: e2 var(b2) = xt x  Correct formula for least squares variance:

Generalized Least Squares et = et1 + t substitute in for et yt = 1 + 2xt + et yt = 1 + 2xt + et1 + t Now we need to get rid of et1 (continued)

(continued) yt = 1 + 2xt + et1 + t yt = 1 + 2xt + et et = yt  12xt lag the errors once et1 = yt1  12xt1 yt = 1 + 2xt + yt1  12xt1 + t (continued)

yt = 1 + 2xt + yt1  12xt1 + t yt  yt1 = 1(1) + 2(xtxt1) + t yt = 1 xt1*+ 2 xt2* + t , * t =2, 3, …, T. yt = yt  yt1 * xt2 = (xtxt1) * x1t* = (1)

yt = yt  yt1 x1t* = (1) xt2 = xt xt1 yt = 1 xt1*+ 2 xt2* + t , * Problems estimating this model with least squares: 1. One observation is used up in creating the transformed (lagged) variables leaving only (T1) observations for estimating the model (Cochrane-Orcutt method drops the first observation). 2. The value of  is not known. We must find some way to estimate it.

(Option) Recovering the 1st Observation Dropping the 1st observation and applying least squares is not the best linear unbiased estimation method. Efficiency is lost because the variance of the error associated with the 1st observation is not equal to that of the other errors. This is a special case of the heteroskedasticity problem except that here all errors are assumed to have equal variance except the 1st error.

Recovering the 1st Observation The 1st observation should fit the original model as: y1 = 1 + 2x1 + e1 with error variance: var(e1) = e2 = 2 /(1-2). We could include this as the 1st observation for our estimation procedure but we must first transform it so that it has the same error variance as the other observations. Note: The other observations all have error variance 2.

y1 = 1 + 2x1 + e1 with error variance: var(e1) = e2 = 2 / (1-2). The other observations all have error variance 2. Given any constant c : var(ce1) = c2 var(e1). If c = 1-2 , then var( 1-2 e1) = (1-2) var(e1). = (1-2) e2 = (1-2) 2 /(1-2) = 2 The transformation 1 = 1-2 e1 has variance 2 .

y1 = 1 + 2x1 + e1 Multiply through by 1-2 to get: The transformed error 1 = 1-2 e1 has variance 2 . This transformed first observation may now be added to the other (T-1) observations to obtain the fully restored set of T observations.

We can summarize these results by saying that, providing  is known, we can find the Best Linear Unbiased Estimator for 1 and 2 by applying least squares to the transformed mode yt = 1 xt1*+ 2 xt2* + t , * t =1, 2, 3, …, T. where the transformed variables are defined by for the first observation, and for the remaining t = 2, 3, …, T observations.

Estimating Unknown  Value If we had values for the et , we could estimate: et = et1 + t First, use least squares to estimate the model: yt = 1 + 2xt + et The residuals from this estimation are: et = yt  b1  b2xt ^

et et-1 et-1 ^ et = yt - b1 - b2xt ^ et = et1 + t ^ = Next, estimate the following by least squares: et = et1 + t ^ The least squares solution is: et et-1 et-1 T t = 2 2 ^ =

Durbin-Watson Test The Durbin-Watson test is by far the most important one for detecting AR(1) errors. It is assumed that the vt are independent random errors with distribution N(0, v2). The assumption of normally distributed random errors is needed to derive the probability distribution of the test statistic used in the Durbin-Watson test.

et et-1 et ^ d= The Durbin-Watson Test statistic, d, is : 2 ^ d= For a null hypothesis of no autocorrelation, we can use H0:  = 0. For an alternative hypothesis we could use H1:  > 0 or H1:  < 0 or H1:   0.

Testing for Autocorrelation The test statistic, d , is approximately related to  as: ^ 0  d  2(1)  4 ^ When  = 0 , the Durbin-Watson statistic is d  2. ^ When  = 1 , the Durbin-Watson statistic is d  0. ^ When  = 1, the Durbin-Watson statistic is d  4. ^ Tables for critical values for d are not always readily available so it is easier to use the p-value that most computer programs provide for d. Reject H0 if p-value < , the significance level.

Test for the first-order autocorrelation Reject H0 Inclusive Do not reject H0 H1:  > 0 d < dL dL < d < dU d > dU H1:  < 0 d > 4  dL 4  dU < d < 4  dL d < 4  dU d > 4  dL 4  dU < d < 4  dL H1:   0 or or dU < d < 4  dU d < dL dL < d < dU Note: The lower and upper bounds (dL and dU) depend on sample size n and the number of explanatory variables k (not include intercept).

 > 0 dU 2 dL 4  < 0 2 4  dU 4  dL 4  > 0  < 0 2 A. Test for Positive Autocorrelation  > 0 No evidence of positive autocorrelation Inclusive dU 2 dL 4 B. Test for Negative Autocorrelation No evidence of negative autocorrelation Inclusive  < 0 2 4  dU 4  dL 4 C. Two-Sided Test for Autocorrelation No evidence of autocorrelation  > 0 Inclusive  < 0 Inclusive 2 4  dL dU 4  dU 4 dL

Prediction with AR(1) Errors When errors are autocorrelated, the previous period error may help us predict next period error. The best predictor, yT+1 , for next period is: yT+1 = 1 + 2xT+1 +  eT ^ ~ where 1 and 2 are generalized least squares estimates and eT is given by: ~ ^ ~ ^ ^ eT = yT  1  2xT

For h periods ahead, the best predictor is: yT+h = 1 + 2xT+h + h eT ^ ~ Assuming |  | < 1, the influence of h eT diminishes the further we go into the future (the larger h becomes). ^ ~