Model specification (identification) We already know about the sample autocorrelation function (SAC): Properties: Not unbiased (since a ratio between two.

Slides:



Advertisements
Similar presentations
Dates for term tests Friday, February 07 Friday, March 07
Advertisements

CHAPTER 13 M ODELING C ONSIDERATIONS AND S TATISTICAL I NFORMATION “All models are wrong; some are useful.”  George E. P. Box Organization of chapter.
Model Building For ARIMA time series
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
Non-stationary data series
Unit Roots & Forecasting
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Building 1. Model Identification
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by means of inference statistical methods.
Economics 310 Lecture 25 Univariate Time-Series Methods of Economic Forecasting Single-equation regression models Simultaneous-equation regression models.
How should these data be modelled?. Identification step: Look at the SAC and SPAC Looks like an AR(1)- process. (Spikes are clearly decreasing in SAC.
Non-Seasonal Box-Jenkins Models
Modeling Cycles By ARMA
SYSTEMS Identification
Prediction and model selection
Financial Time Series CS3. Financial Time Series.
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
Financial Econometrics
Economics 20 - Prof. Anderson
Modern methods The classical approach: MethodProsCons Time series regression Easy to implement Fairly easy to interpret Covariates may be added (normalization)
Non-Seasonal Box-Jenkins Models
ARMA models Gloria González-Rivera University of California, Riverside
Time Series Forecasting (Part II)
Time Series Analysis.
STAT 497 LECTURE NOTES 2.
STAT 497 LECTURE NOTES 4 MODEL INDETIFICATION AND NON- STATIONARY TIME SERIES MODELS 1.
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 8: Estimation & Diagnostic Checking in Box-Jenkins.
Linear Stationary Processes. ARMA models. This lecture introduces the basic linear models for stationary processes. Considering only stationary processes.
TIME SERIES ANALYSIS Time Domain Models: Red Noise; AR and ARMA models LECTURE 7 Supplementary Readings: Wilks, chapters 8.
MODEL INDETIFICATION AND NON-STATIONARY TIME SERIES MODELS
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Byron Gangnes Econ 427 lecture 11 slides Moving Average Models.
K. Ensor, STAT Spring 2005 Model selection/diagnostics Akaike’s Information Criterion (AIC) –A measure of fit plus a penalty term for the number.
The Properties of Time Series: Lecture 4 Previously introduced AR(1) model X t = φX t-1 + u t (1) (a) White Noise (stationary/no unit root) X t = u t i.e.
Big Data at Home Depot KSU – Big Data Survey Course Steve Einbender Advanced Analytics Architect.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
How do we identify non-stationary processes? (A) Informal methods Thomas 14.1 Plot time series Correlogram (B) Formal Methods Statistical test for stationarity.
Time Series Basics Fin250f: Lecture 8.1 Spring 2010 Reading: Brooks, chapter
3.Analysis of asset price dynamics 3.1Introduction Price – continuous function yet sampled discretely (usually - equal spacing). Stochastic process – has.
MULTIVARIATE TIME SERIES & FORECASTING 1. 2 : autocovariance function of the individual time series.
Auto Regressive, Integrated, Moving Average Box-Jenkins models A stationary times series can be modelled on basis of the serial correlations in it. A non-stationary.
Time Series Analysis Lecture 11
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
Forecasting (prediction) limits Example Linear deterministic trend estimated by least-squares Note! The average of the numbers 1, 2, …, t is.
Autoregressive (AR) Spectral Estimation
Review and Summary Box-Jenkins models Stationary Time series AR(p), MA(q), ARMA(p,q)
The Box-Jenkins (ARIMA) Methodology
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
Univariate Time series - 2 Methods of Economic Investigation Lecture 19.
Geology 6600/7600 Signal Analysis 04 Sep 2014 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
Introduction to stochastic processes
Stationarity and Unit Root Testing Dr. Thomas Kigabo RUSUHUZWA.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Adv DSP Spring-2015 Lecture#11 Spectrum Estimation Parametric Methods.
Review of Unit Root Testing D. A. Dickey North Carolina State University (Previously presented at Purdue Econ Dept.)
Analysis of Financial Data Spring 2012 Lecture 4: Time Series Models - 1 Priyantha Wijayatunga Department of Statistics, Umeå University
CH 5: Multivariate Methods
PSG College of Technology
Model Building For ARIMA time series
Applied Econometric Time-Series Data Analysis
ARMA models 2012 International Finance CYCU
Cross-validation for the selection of statistical models
Unit Root & Augmented Dickey-Fuller (ADF) Test
Chapter 8. Model Identification
Vector AutoRegression models (VARs)
The Spectral Representation of Stationary Time Series
Time Series introduction in R - Iñaki Puigdollers
CH2 Time series.
BOX JENKINS (ARIMA) METHODOLOGY
Presentation transcript:

Model specification (identification) We already know about the sample autocorrelation function (SAC): Properties: Not unbiased (since a ratio between two random variables) Bias decreases with n Variance complicated, common to use general large-sample results

Large-sample results (asymptotics): For large n the random vector has an approximate multivariate normal distribution with zero mean vector and covariance matrix ( c ij ) where This gives that  Var (r k )  0 as n  does not diminish as n 

Hence, the distribution of r k will depend on the correlation structure of Y t and accordingly on the model behind (i.e. if it is and AR(1), an ARMA(2,1) etc.) For an AR(1), i.e. Y t =   Y t – 1 + e t For an MA(q) i.e. not dependent on k after the qth lag i.e. not dependent on k for large lags For white noise

Partial autocorrelation function Describes the “specific” part of the correlation between Y t and Y t – k that is not due to successive serial correlations between the variables Y t – 1, Y t – 2, …, Y t – k. Partial correlations are used for other types of data as well (for instance in linear models of cross-sectional data. Patterns For an AR(p)-process,  k cuts off after lag p (i.e. the same type of behaviour like  k has for an MA(q)-process For an MA(q)-process  k shows approximately the same pattern as does  k for an AR(p)-process

Estimation from data, Sample Partial Autocorrelation function (SPAC): No explicit formula, estimation has to be made recursively Properties of SPAC: More involved, but for an AR(p)-process SPAC-values for lags greater than p are approximately normally distributed with zero mean and variance 1/n

Extended Autocorrelation function (EACF) One (of several) tool to improve the choice of orders of ARMA(p, q)-processes. Very clear as a theoretical function, but noisy when estimated on series not too long. AR, MA or ARMA?

No pattern at all?

EACF table for Y AR/MA o o o o o o o o o o o o o o 1 o o o o o o o o o o o o o o 2 x o o o o o o o o o o o o o 3 o o x o o o o o o o o o o o 4 x o o o o o o o o o o o o o 5 x x o o o o o o o o o o o o 6 o x o o o x o o o o o o o o 7 o o o o o x o o o o o o o o ARMA(0,0) or ARMA(1,0)? True process: Y t =  Y t – 1 + e t – 0.1  e t – 1

Model selection from more analytical tools Dickey-Fuller Unit-Root test H 0 : The process Y t is difference non-stationary (  Y t is stationary) H a : The process Y t is stationary Augmented Dickey-Fuller test statistic (ADF): If  =1 (difference non-stationary) 

Fit the model and test H 0 : a = 0 (difference non-stationary) vs. H a : a < 0 (stationary) using the test statistic However, not t-distributed under H 0. Another sampling distributions has been derived and tables (programmed in R)

Akaike’s criteria For an ARMA(p,q)-process, let k = p + q + 1 and find the values of the parameters p, q,  1, …,  p,  1, …,  q of the model that minimizes –2  log{max L(p, q,  1, …,  p,  1, …,  q )} + 2  k AIC [ Akaike’s Information Criterion]. Works well when the true process have (at least one) infinite order –2  log{max L(p, q,  1, …,  p,  1, …,  q )} + k  log(n) BIC [ (Schwarz) Bayesian Information Criterion]. Works well when we “know” that the true process is a finite-order ARMA(p,q)