Part II – TIME SERIES ANALYSIS C4 Autocorrelation Analysis © Angel A. Juan & Carles Serrat - UPC 2007/2008.

Slides:



Advertisements
Similar presentations
SMA 6304 / MIT / MIT Manufacturing Systems Lecture 11: Forecasting Lecturer: Prof. Duane S. Boning Copyright 2003 © Duane S. Boning. 1.
Advertisements

Autocorrelation Functions and ARIMA Modelling
Part II – TIME SERIES ANALYSIS C3 Exponential Smoothing Methods © Angel A. Juan & Carles Serrat - UPC 2007/2008.
2013/12/10.  The Kendall’s tau correlation is another non- parametric correlation coefficient  Let x 1, …, x n be a sample for random variable x and.
Using SAS for Time Series Data
Inference for Regression
Model Building For ARIMA time series
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
STATIONARY AND NONSTATIONARY TIME SERIES
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Analysis Materials for this lecture Lecture 5 Lags.XLS Lecture 5 Stationarity.XLS Read Chapter 15 pages Read Chapter 16 Section 15.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Part II – TIME SERIES ANALYSIS C1 Introduction to TSA © Angel A. Juan & Carles Serrat - UPC 2007/2008 "If we could first know where we are, then whither.
IB Math Studies – Topic 6 Statistics.
Economics 310 Lecture 25 Univariate Time-Series Methods of Economic Forecasting Single-equation regression models Simultaneous-equation regression models.
How should these data be modelled?. Identification step: Look at the SAC and SPAC Looks like an AR(1)- process. (Spikes are clearly decreasing in SAC.
Correlation Chapter 9.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Designing Experiments In designing experiments we: Manipulate the independent.
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
Financial Econometrics
Modern methods The classical approach: MethodProsCons Time series regression Easy to implement Fairly easy to interpret Covariates may be added (normalization)
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Discovering and Describing Relationships
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Autocorrelation Lecture 18 Lecture 18.
1 King Abdulaziz University Faculty of Engineering Industrial Engineering Dept. IE 436 Dynamic Forecasting.
Stat 112: Lecture 16 Notes Finish Chapter 6: –Influential Points for Multiple Regression (Section 6.7) –Assessing the Independence Assumptions and Remedies.
Business Statistics - QBM117 Statistical inference for regression.
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
BOX JENKINS METHODOLOGY
AR- MA- och ARMA-.
Inference for regression - Simple linear regression
TIME SERIES by H.V.S. DE SILVA DEPARTMENT OF MATHEMATICS
STAT 497 LECTURE NOTES 2.
The Examination of Residuals. The residuals are defined as the n differences : where is an observation and is the corresponding fitted value obtained.
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
#1 EC 485: Time Series Analysis in a Nut Shell. #2 Data Preparation: 1)Plot data and examine for stationarity 2)Examine ACF for stationarity 3)If not.
Tutorial for solution of Assignment week 39 “A. Time series without seasonal variation Use the data in the file 'dollar.txt'. “
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Autoregressive Integrated Moving Average (ARIMA) Popularly known as the Box-Jenkins methodology.
John G. Zhang, Ph.D. Harper College
Introducing ITSM 2000 By: Amir Heshmaty Far. S EVERAL FUNCTION IN ITSM to analyze and display the properties of time series data to compute and display.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Geo479/579: Geostatistics Ch4. Spatial Description.
STAT 497 LECTURE NOTE 9 DIAGNOSTIC CHECKS 1. After identifying and estimating a time series model, the goodness-of-fit of the model and validity of the.
Time Series Analysis Lecture 11
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Correlation & Regression Analysis
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Example x y We wish to check for a non zero correlation.
The Box-Jenkins (ARIMA) Methodology
Seasonal ARIMA FPP Chapter 8.
2/25/ lecture 121 STATS 330: Lecture 12. 2/25/ lecture 122 Diagnostics 4 Aim of today’s lecture To discuss diagnostics for independence.
Ch16: Time Series 24 Nov 2011 BUSI275 Dr. Sean Ho HW8 due tonight Please download: 22-TheFed.xls 22-TheFed.xls.
Analysis of financial data Anders Lundquist Spring 2010.
MODEL DIAGNOSTICS By Eni Sumarminingsih, Ssi, MM.
Chapter 15 Multiple Regression Model Building
Covariance, stationarity & some useful operators
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
Lecture 8 ARIMA Forecasting II
Chapter 6: Autoregressive Integrated Moving Average (ARIMA) Models
Elementary Statistics
STAT 497 LECTURE NOTE 9 DIAGNOSTIC CHECKS.
STATIONARY AND NONSTATIONARY TIME SERIES
BOX JENKINS (ARIMA) METHODOLOGY
Chap 7: Seasonal ARIMA Models
Presentation transcript:

Part II – TIME SERIES ANALYSIS C4 Autocorrelation Analysis © Angel A. Juan & Carles Serrat - UPC 2007/2008

2.4.1: Lagged Values & Dependencies  Often in time series you will want to compare the value observed at one time point to a value observed one or more time points earlier. Such prior values are known as lagged values.  File: RIVERC.MTW  Stat > Time Series > Lag… It is frequent to detect a strong relationship (correlation) between the original data and some of the lagged data sets. This is specially common for the cases lag = 1 and lag = k, being k is the size of the seasonal component (k = 24 in this example). In this example, it is clear that each observation is most similar (closest) to the adjacent observation (lag = 1); in addition, there is a recurring seasonal pattern, i.e., each observation is also similar to the same observation 24 hours earlier (lag = 24). You can calculate lagged values by letting the values in rows of the lagged column be equal to values one or more rows above in the unlagged column.

2.4.2: Autocorrelation Function ACF (1/2)  Idea: If there is some pattern in how the values of your time series change from observations to observation, you could use it to your advantage.  The correlation between the original time series values and the corresponding k-lagged values is called autocorrelation of order k.  The Autocorrelation Function (ACF) provides the correlation between the serial correlation coefficients for consecutive lags.  Correlograms display graphically the ACF.  The ACF can be misleading for a series with unstable variance, so it might first be necessary to transform for a constant variance before using the ACF. For instance, a below-average value at time t means that it is more likely the series will be high at time t+1, or viceversa. For this TS, autocorrelation of order 2 (correlation between original TS data and 2-lagged data) is about The (1-α) confidence boundary indicates how low or high the correlation needs to be for significance at the α level. Autocorrelations that lie outside this range are statistically significant.

2.4.2: Autocorrelation Function ACF (2/2)  File: RIVERC.MTW  Stat > Time Series > Autocorrelation… The autocorrelation plot resembles a sine pattern. This suggest that temperatures close in time are strongly and positively correlated, while temperatures about 12 hours apart are highly negatively correlated. The large positive autocorrelation at lag 24 is indicative of the 24-hour seasonal component. You can use the Ljung-Box Q (LBQ) statistic to test the null hypothesis that the autocorrelations for all lags up to lag k equal zero (the LBQ is Chi-Square distributed with df = k). Alternatively, when using α = 0.05, you can simply check whether the ACF value is beyond the significance limits.

2.4.3: Serial Dependency & Differencing  Note: Autocorrelations for consecutive lags are formally dependent.  Example: If the first element is closely related to the second, and the second to the third, then the first element must also be somewhat related to the third one, etc.  Implication: The pattern of serial dependencies can change considerably after removing the first order autocorrelation.  Why to remove serial dependency?: 1.To identify the hidden nature of seasonal dependencies in the series. 2.To make the series stationary which is necessary for ARIMA and other techniques.  How to remove it?: Serial dependency for a particular lag of k can be removed by differencing the series, that is converting each element i of the series into its difference from the element i-k.  File: RIVERC.MTW  Stat > Time Series > Differences… Recall that the shape of your TS plot indicates whether the TS is stationary or nonstationary. Consider taking differences when the TS plot indicates a nonstationary time series. You can then use the resulting differenced data as your time series, plotting the differenced TS to determine if it is stationary.

2.4.4: Partial Autocorr. Function PACF (1/2)  Another useful method to examine serial dependencies is to examine the Partial Autocorrelation Function (PACF), an extension of autocorrelation where the dependence on the intermediate elements (those within the lag) is removed.  For time series data, ACF and PACF measure the degree of relationship between observations k time periods, or lags, apart. These plots provide valuable information to help you identify an appropriate ARIMA model.  In a sense, the partial autocorrelation provides a “cleaner” picture of serial dependencies for individual lags. At the lag of 1 (when there are no intermediate elements within the lag), the partial autocorrelation is equivalent to autocorrelation

2.4.4: Partial Autocorr. Function PACF (2/2)  File: RIVERC.MTW  Stat > Time Series > Partial Autocorrelation… The graph shows two large partial autocorrelations at lags 1 and 2.