Download presentation
Presentation is loading. Please wait.
Published byShannon Carroll Modified over 9 years ago
1
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes
2
9.1 Introduction 9.2 Lags in the Error Term: Autocorrelation 9.3 Estimating an AR(1) Error Model 9.4 Testing for Autocorrelation 9.5 An Introduction to Forecasting: Autoregressive Models 9.6 Finite Distributed Lags 9.7 Autoregressive Distributed Lag Models
3
Figure 9.1
4
The model so far assumes that the observations are not correlated with one another. This is believable if one has drawn a random sample, but less likely if one has drawn observations sequentially in time Time series observations, which are drawn at regular intervals, usually embody a structure where time is an important component. Principles of Econometrics, 3rd Edition
5
If we cannot completely model this structure in the regression function itself, then the remainder spills over into the unobserved component of the statistical model (its error) and this causes the errors be correlated with one another. Principles of Econometrics, 3rd Edition
6
Three ways to view the dynamics:
7
Figure 9.2(a) Time Series of a Stationary Variable Assume Stationarity For now
8
Figure 9.2(b) Time Series of a Nonstationary Variable that is ‘Slow Turning’ or ‘Wandering’
9
Figure 9.2(c) Time Series of a Nonstationary Variable that ‘Trends’
10
9.2.1 Area Response Model for Sugar Cane
11
Assume also
12
OK OK, constant Not zero!!!
13
Aka autocorrelation coefficient
15
Figure 9.3 Least Squares Residuals Plotted Against Time
16
Constant variance Null expectation of x and y
17
The existence of AR(1) errors implies: The least squares estimator is still a linear and unbiased estimator, but it is no longer best. There is another estimator with a smaller variance. The standard errors usually computed for the least squares estimator are incorrect. Confidence intervals and hypothesis tests that use these standard errors may be misleading.
18
As with heteroskedastic errors, you can salvage OLS when your data are autocorrelated. Now you can use an estimator of standard errors that is robust to both heteroskedasticity and autocorrelation proposed by Newey and West. This estimator is sometimes called HAC, which stands for heteroskedasticity autocorrelated consistent.
19
HAC is not as automatic as the heteroskedasticity consistent (HC) estimator. With autocorrelation you have to specify how far away in time the autocorrelation is likely to be significant The autocorrelated errors over the chosen time window are averaged in the computation of the HAC standard errors; you have to specify how many periods over which to average and how much weight to assign each residual in that average. That weighted average is called a kernel and the number of errors to average is called bandwidth.
20
Usually, you choose a method of averaging (Bartlett kernel or Parzen kernel) and a bandwidth (nw1, nw2 or some integer). Often your software defaults to the Bartlett kernel and a bandwidth computed based on the sample size, N. Trade-off: Larger bandwidths reduce bias (good) as well as precision (bad). Smaller bandwidths exclude more relevant autocorrelations (and hence have more bias), but use more observations to increase precision (smaller variance). Choose a bandwidth large enough to contain the largest autocorrelations. The choice will ultimately depend on the frequency of observation and the length of time it takes for your system to adjust to shocks.
21
Sugar cane example The two sets of standard errors, along with the estimated equation are: The 95% confidence intervals for β 2 are:
23
Now this transformed nonlinear model has errors that are uncorrelated over time
24
It can be shown that nonlinear least squares estimation of (9.24) is equivalent to using an iterative generalized least squares estimator called the Cochrane-Orcutt procedure. Details are provided in Appendix 9A.
25
The nonlinear least squares estimator only requires that the errors be stable (not necessarily stationary). Other methods commonly used make stronger demands on the data, namely that the errors be covariance stationary. Furthermore, the nonlinear least squares estimator gives you an unconditional estimate of the autocorrelation parameter,, and yields a simple t-test of the hypothesis of no serial correlation. Monte Carlo studies show that it performs well in small samples as well.
26
But nonlinear least squares requires more computational power than linear estimation, though this is not much of a constraint these days. Nonlinear least squares (and other nonlinear estimators) use numerical methods rather than analytical ones to find the minimum of your sum of squared errors objective function. The routines that do this are iterative.
27
We should then test the above restrictions
28
9.4.1Residual Correlogram
29
9.4.1Residual Correlogram (more practically…)
30
Figure 9.4 Correlogram for Least Squares Residuals from Sugar Cane Example
31
For this nonlinear model, then, the residuals should be uncorrelated
32
Figure 9.5 Correlogram for Nonlinear Least Squares Residuals from Sugar Cane Example
33
The other way to determine whether or not your residuals are autocorrelated is to use an LM (Lagrange multiplier) test. For autocorrelation, this test is based on an auxiliary regression where lagged OLS residuals are added to the original regression equation. If the coefficient on the lagged residual is significant then you conclude that the model is autocorrelated.
34
We can derive the auxiliary regression for the LM test:
35
Centered around zero, so the power of this test now would come from The lagged error, which is what we care about…
36
This was the Breusch-Godfrey LM test for autocorrelation and it is distributed chi-sq In STATA: regress la lp estat bgodfrey
37
Adding lagged values of y can serve to eliminate the error autocorrelation In general, p should be large enough so that vt is white noise
38
Figure 9.6 Correlogram for Least Squares Residuals from AR(3) Model for Inflation Helps decide How many lags To include in the model
44
This model is just a generalization of the ones previously discussed. In this model you include lags of the dependent variable (autoregressive) and the contemporaneous and lagged values of independent variables as regressors (distributed lags). The acronym is ARDL(p,q) where p is the maximum distributed lag and q is the maximum autoregressive lag
49
Figure 9.7 Correlogram for Least Squares Residuals from Finite Distributed Lag Model
51
Figure 9.8 Correlogram for Least Squares Residuals from Autoregressive Distributed Lag Model
53
Figure 9.9 Distributed Lag Weights for Autoregressive Distributed Lag Model
54
Slide 9-54 Principles of Econometrics, 3rd Edition autocorrelation autoregressive distributed lag models autoregressive error autoregressive model correlogram delay multiplier distributed lag weight dynamic models finite distributed lag forecast error forecasting HAC standard errors impact multiplier infinite distributed lag interim multiplier lag length lagged dependent variable LM test nonlinear least squares sample autocorrelation function standard error of forecast error total multiplier form of LM test
55
Slide 9-55 Principles of Econometrics, 3rd Edition
56
Slide 9-56 Principles of Econometrics, 3rd Edition (9A.2) (9A.1)
57
Slide 9-57 Principles of Econometrics, 3rd Edition (9A.4) (9A.3)
58
Slide 9-58 Principles of Econometrics, 3rd Edition (9A.5) (9A.6)
59
Slide 9-59 Principles of Econometrics, 3rd Edition
60
Slide 9-60 Principles of Econometrics, 3rd Edition (9B.1)
61
Slide 9-61 Principles of Econometrics, 3rd Edition (9B.2)
62
Slide 9-62 Principles of Econometrics, 3rd Edition (9B.3)
63
Figure 9A.1: Principles of Econometrics, 3rd Edition Slide 9-63
64
Figure 9A.2: Principles of Econometrics, 3rd Edition Slide 9-64
65
The Durbin-Watson bounds test. if the test is inconclusive. Principles of Econometrics, 3rd Edition Slide 9-65
66
Slide 9-66 Principles of Econometrics, 3rd Edition
67
Slide 9-67 Principles of Econometrics, 3rd Edition (9C.2) (9C.1)
68
Slide 9-68 Principles of Econometrics, 3rd Edition (9C.3)
69
Slide 9-69 Principles of Econometrics, 3rd Edition (9C.4)
70
Slide 9-70 Principles of Econometrics, 3rd Edition
71
Slide 9-71 Principles of Econometrics, 3rd Edition (9C.6) (9C.5)
72
Slide 9-72 Principles of Econometrics, 3rd Edition (9D.2) (9D.1)
73
Figure 9A.3: Exponential Smoothing Forecasts for two alternative values of α Principles of Econometrics, 3rd Edition Slide 9-73
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.