Some more issues of time series analysis Time series regression with modelling of error terms In a time series regression model the error terms are tentatively.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Managerial Economics in a Global Economy
Time series. Characteristics Non-independent observations (correlations structure) Systematic variation within a year (seasonal effects) Long-term increasing.
Hypothesis Testing Steps in Hypothesis Testing:
CH 27. * Data were collected on 208 boys and 206 girls. Parents reported the month of the baby’s birth and age (in weeks) at which their child first crawled.
Objectives (BPS chapter 24)
How should these data be modelled?. Identification step: Look at the SAC and SPAC Looks like an AR(1)- process. (Spikes are clearly decreasing in SAC.
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Lecture 3: Chi-Sqaure, correlation and your dissertation proposal Non-parametric data: the Chi-Square test Statistical correlation and regression: parametric.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Exercise 7.1 a. Find and report the four seasonal factors for quarter 1, 2, 3 and 4 sn1 = 1.191, sn2 = 1.521, sn3 = 0.804, sn4 = b. What is the.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
REGRESSION MODEL ASSUMPTIONS. The Regression Model We have hypothesized that: y =  0 +  1 x +  | | + | | So far we focused on the regression part –
Stat 217 – Day 25 Regression. Last Time - ANOVA When?  Comparing 2 or means (one categorical and one quantitative variable) Research question  Null.
REGRESSION AND CORRELATION
Part 18: Regression Modeling 18-1/44 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics.
15-1 Introduction Most of the hypothesis-testing and confidence interval procedures discussed in previous chapters are based on the assumption that.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Correlation and Regression Analysis
Descriptive measures of the strength of a linear association r-squared and the (Pearson) correlation coefficient r.
Linear Regression/Correlation
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
AR- MA- och ARMA-.
Inference for regression - Simple linear regression
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Regression Method.
CPE 619 Simple Linear Regression Models Aleksandar Milenković The LaCASA Laboratory Electrical and Computer Engineering Department The University of Alabama.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
NONPARAMETRIC STATISTICS
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
1 Trend Analysis Step vs. monotonic trends; approaches to trend testing; trend tests with and without exogeneous variables; dealing with seasonality; Introduction.
CHAPTER 14 MULTIPLE REGRESSION
INTRODUCTORY LINEAR REGRESSION SIMPLE LINEAR REGRESSION - Curve fitting - Inferences about estimated parameter - Adequacy of the models - Linear.
Chapter 12: Linear Regression 1. Introduction Regression analysis and Analysis of variance are the two most widely used statistical procedures. Regression.
Introduction to Linear Regression
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Tutorial for solution of Assignment week 39 “A. Time series without seasonal variation Use the data in the file 'dollar.txt'. “
Biostatistics, statistical software VII. Non-parametric tests: Wilcoxon’s signed rank test, Mann-Whitney U-test, Kruskal- Wallis test, Spearman’ rank correlation.
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
Lack of Fit (LOF) Test A formal F test for checking whether a specific type of regression function adequately fits the data.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Lecture 10 Chapter 23. Inference for regression. Objectives (PSLS Chapter 23) Inference for regression (NHST Regression Inference Award)[B level award]
Simple linear regression Tron Anders Moger
1 Experimental Statistics - week 12 Chapter 12: Multiple Regression Chapter 13: Variable Selection Model Checking.
STAT 497 LECTURE NOTE 9 DIAGNOSTIC CHECKS 1. After identifying and estimating a time series model, the goodness-of-fit of the model and validity of the.
Forecasting (prediction) limits Example Linear deterministic trend estimated by least-squares Note! The average of the numbers 1, 2, …, t is.
Correlation & Regression Analysis
Chapter 8: Simple Linear Regression Yang Zhenlin.
Residual Analysis Purposes –Examine Functional Form (Linear vs. Non- Linear Model) –Evaluate Violations of Assumptions Graphical Analysis of Residuals.
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
2/25/ lecture 121 STATS 330: Lecture 12. 2/25/ lecture 122 Diagnostics 4 Aim of today’s lecture To discuss diagnostics for independence.
Maths Study Centre CB Open 11am – 5pm Semester Weekdays
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Regression Analysis Presentation 13. Regression In Chapter 15, we looked at associations between two categorical variables. We will now focus on relationships.
Chapter 15 Inference for Regression. How is this similar to what we have done in the past few chapters?  We have been using statistics to estimate parameters.
Simple Linear Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Some more issues of time series analysis Time series regression with modelling of error terms In a time series regression model the error terms are tentatively assumed to be independent and identically distributed. Is this wise?

Performing e.g. the Durbin-Watson test we may quite easily answer the question whether they are or not. What if D-W gives evidence of serial correlation in the error terms? Apply an AR(p) model to the error terms at the same time as the rest of the model is fitted. Standard procedure: Study the residuals from an ordinary regression fit Identify which order p of the AR-model that may be the most appropriate for the error terms. Make the fit of the combined regression-AR-model

Estimation can no longer be done using ordinary least-squares. Instead the conditional least-squares method is used. Procedures are not curretly available in Minitab, but in more comprehensive computer packages such as SAS and SPSS. Example Consider again the Hjälmaren month data set (that is used in assignments for weeks 36, 39 and 41)

Minitab output from an ordinary time series regression: The regression equation is Discharge.m = Time.m Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Predictor Coef SE Coef T P Constant Time.m Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov S = R-Sq = 18.8% R-Sq(adj) = 18.1%

Residual plots Residuals seem to follow an AR- model with order 1 or 2

SPSS output of a regression analysis with error term modelled as AR(1) FINAL PARAMETERS: Number of residuals 1284 Standard error Log likelihood AIC SBC Variables in the Model: B SEB T-RATIO APPROX. PROB. AR JAN FEB MAR APR MAY JUN JUL AUG SEP OCT NOV TIME CONSTANT Variance of pure error term smaller than variance of error term in ordinary regression!

Non-parametric tests for trend All models so far taken up in the course are parametric models. Parametric models assume a specific probability distribution is governing the obtained observations (i.e. the normal distribution) and The population mean value of each observation can be expressed in terms of the parameters of the model. What if we cannot specify this probability distribution?

Least-squares fitting of time series regression models can still be done, but none of the significance tests are valid  We cannot test for the presence of a trend (nor for the presence of seasonal variation) Classical decomposition is still possible but they have no significance tests built-in (they are all descriptive analysis tools) Conditional least-squares estimation in ARIMA models are not valid as they emerge from the assumption that the observations are normally distributed. As a consequence the significant tests are not valid.

The Mann-Kendall test for a monotonic trend Example: Look again at the data set of sales values from lecture 3, but with restriction to the years Year Sales values Could there be a trend in data?

If there is a trend, we do not assume that it has a specific functionalform, such as linear or quadratic, just assume it is monotonic, i.e. decreasing or increasing. In this case it would be a decreasing trend. The sign function:

Now define the Mann-Kendall test statistic as i.e. the statistic is a sum of +1:s, –1:s and 0:s depending on whether y j is higher than, lower than or equal to y i for each pair of time points (i, j : i < j). Large positive values of T would then be consistent with an upward trend Large negative values of T would be consistent with a downward trend Values around 0 of T would be consistent with no trend

For the current data set: Now, is T = – 43 enough negatively large to show evidence for a trend?

The non-parametric initial “fashion”: Calculate all possible values of T by letting each difference y j – y i, i < j have in order the signs –1, 0 and 1. (Put these values in ascending order ) For the test of H 0 : No trend vs. H A : Negative monotonic trend at the level of significance , calculate the (100  )th percentile of the (ordered) values  T  If the observed T is < T  reject H 0, otherwise “accept” H 0 If a fairly long length of the time series this procedure is quite tedious.

Approximate solution: The variance of T can be shown to be where n is the length of the time series, g is the number of so-called ties (ties means values that have duplicates) and t p is the number of duplicates for tie p. Then for fairly large n

For the current time series of sales values: n = 11 g = 3 (the values 143, 145 and 151 have each two duplicates)  t 1 = t 2 = t 3 = 2  Var (T ) = (1/18)  (11  10  27 – (3  2  1  9)) = 162  P-value is Thus H 0 may be rejected at any reasonable level of significance

For time series with seasonal variation, Hirsch & Sclack has developed a modification of the Mann-Kendall test with test statistic where T k is the Mann-Kendall test statistic for the time series consisting of values from season k only (e.g. for montly data we consider the series of January values, the series of February values etc.) Expressions for the variance of TS can be derived and analogously to the Mann- Kendall test