Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.

Slides:



Advertisements
Similar presentations
SJS SDI_21 Design of Statistical Investigations Stephen Senn 2 Background Stats.
Advertisements

Autocorrelation Functions and ARIMA Modelling
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Hypothesis Testing Steps in Hypothesis Testing:
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Model Building For ARIMA time series
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Model specification (identification) We already know about the sample autocorrelation function (SAC): Properties: Not unbiased (since a ratio between two.
Linear regression models
R. Werner Solar Terrestrial Influences Institute - BAS Time Series Analysis by means of inference statistical methods.
SOLVED EXAMPLES.
STAT 497 LECTURE NOTES 8 ESTIMATION.
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
STAT 497 APPLIED TIME SERIES ANALYSIS
How should these data be modelled?. Identification step: Look at the SAC and SPAC Looks like an AR(1)- process. (Spikes are clearly decreasing in SAC.
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
Chapter 10 Simple Regression.
9. SIMPLE LINEAR REGESSION AND CORRELATION
Point estimation, interval estimation
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Maximum-Likelihood estimation Consider as usual a random sample x = x 1, …, x n from a distribution with p.d.f. f (x;  ) (and c.d.f. F(x;  ) ) The maximum.
Chapter 11 Multiple Regression.
OMS 201 Review. Range The range of a data set is the difference between the largest and smallest data values. It is the simplest measure of dispersion.
Simple Linear Regression Analysis
July 3, Department of Computer and Information Science (IDA) Linköpings universitet, Sweden Minimal sufficient statistic.
Basics of regression analysis
Analysis of Individual Variables Descriptive – –Measures of Central Tendency Mean – Average score of distribution (1 st moment) Median – Middle score (50.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Business Statistics - QBM117 Statistical inference for regression.
Maximum likelihood (ML)
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Statistical Intervals Based on a Single Sample.
Introduction to Regression Analysis, Chapter 13,
Chapter 12 Section 1 Inference for Linear Regression.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
AR- MA- och ARMA-.
ARMA models Gloria González-Rivera University of California, Riverside
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
STAT 497 LECTURE NOTES 2.
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
1 G Lect 10a G Lecture 10a Revisited Example: Okazaki’s inferences from a survey Inferences on correlation Correlation: Power and effect.
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 8: Estimation & Diagnostic Checking in Box-Jenkins.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
K. Ensor, STAT Spring 2005 Model selection/diagnostics Akaike’s Information Criterion (AIC) –A measure of fit plus a penalty term for the number.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Forecasting (prediction) limits Example Linear deterministic trend estimated by least-squares Note! The average of the numbers 1, 2, …, t is.
Chapter 8: Simple Linear Regression Yang Zhenlin.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Review and Summary Box-Jenkins models Stationary Time series AR(p), MA(q), ARMA(p,q)
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
CORRELATION-REGULATION ANALYSIS Томский политехнический университет.
MODEL DIAGNOSTICS By Eni Sumarminingsih, Ssi, MM.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Standard Errors Beside reporting a value of a point estimate we should consider some indication of its precision. For this we usually quote standard error.
Dynamic Models, Autocorrelation and Forecasting
Model Building For ARIMA time series
Statistical Methods For Engineers
Machine Learning Week 4.
Simple Linear Regression
Lecture # 2 MATHEMATICAL STATISTICS
CH2 Time series.
Presentation transcript:

Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population moments with the corresponding sample moments: Trivial MM estimates are estimates of the population mean (  ) and the population variance (  2 ). The benefit of the method is that the equations render possibilities to estimate other parameters.

Mixed moments Moments can be raw (e.g. the mean) or central (e.g. the variance). There are also mixed moments like the covariance and the correlation (which are also central). MM-estimation of parameters in ARMA-models is made by equating the autocorrelation function with the sample autocorrelation function for a sufficient number of lags. For AR-models: Replace  k by r k in the Yule-Walker equations For MA-models: Use developed relationships between  k and the parameters  1, …,  q and replace  k by r k in these.  Leads quickly to complicated equations with no unique solution. Mixed ARMA: As complicated as the MA-case

Example of formulas AR(1): AR(2):

MA(1): Only one solution at a time gives an invertible MA-process

The parameter  e 2 : Set  0 = s 2 

Example Simulated from the model > ar(yar1,method="yw") Call: ar(x = yar1, method = "yw") Coefficients: Order selected 1 sigma^2 estimated as “Yule-Walker (leads to MM-estimates)

Least-squares estimation Ordinary least-squares Find the parameter values p 1, …, p m that minimise the square sum where X stands for an array of auxiliary variables that are used as predictors for Y. Autoregressive models The counterpart of S (p 1, …, p m ) is Here, we take into account the possibility of a mean different from zero.

Now, the estimation can be made in two steps: 1)Estimate  by 2)Find the values of  1, …,  p that minimises The estimation of the slope parameters thus becomes conditional on the estimation of the mean. The square sum S c is therefore referred to as the conditional sum-of-squares function. The resulting estimates become very close to the MM-estimate for moderately long series.

Moving average models More tricky, since each observed value is assumed to depend on unobservable white-noise terms (and a mean): As for the AR-case, first estimate the mean and then estimate the slope parameters conditionally on the estimated mean, i.e. For an invertible MA-process we may write

The square sum to be minimized is then generally Problems: The representation is infinite, but we only have a finite number of observed values S c is a nonlinear function of the parameters  1, …,  q  Numerical solution is needed Compute e t recursively using the observed values Y 1, …, Y n and setting e 0 = e –1 = … = e –q = 0 : for a certain set of values  1, …,  q Numerical algorithms used to find the set that minimizes

Mixed Autoregressive and Moving average models Least-squares estimation is applied analogously to pure MA-models. e t –values are recursively calculated setting e p = e p – 1 = … = e p + 1 – q = 0 Least-squares generally works well for long series For moderately long series the initializing with e-values set to zero may have too much influence on the estimates.

Maximum-Likelihood-estimation (MLE) For a set of observations y 1, …, y n the likelihood function (of the parameters) is a function proportional to the joint density (or probability mass) function of the corresponding random variables Y 1, …, Y n evaluated at those observations: For a times series such a function is not the product of the marginal densities/probability mass functions. We must assume a probability distribution for the random variables. For time series it is common to assume that the white noise is normally distributed, i.e.

with known joint density function For the AR(1) case we can use that the model defines a linear transformation to form Y 2, …, Y n from Y 1, …, Y n–1 and e 2, …, e n This transformation has Jacobian = 1 which simplifies the derivation of the joint density for Y 2, …, Y n given Y 1 to

Now Y 1 should be normally distributed with mean  and variance  e 2 /(1–  2 ) according to the derived properties and the assumption of normally distributed e. Hence the likelihood function becomes and the MLEs of the parameters .  and  e 2 are found as the values that maximises L

Compromise between MLE and Conditional least-squares: Unconditional least-squares estimates of  and  are found by minimising The likelihood function can be put up for ant ARMA-model, however it is more involved for models more complex than AR(1). The estimation need (with a few exceptions) to be carried out numerically

Properties of the estimates Maximum-Likelihood estimators has a well-established asymptotic theory: Hence, by deriving large-sample expressions for the variances of the point estimates, these can be used to make inference about the parameters (tests and confidence intervals) See the textbook

Model diagnostics Upon estimation of a model, its residuals should be checked as usual. Residuals should be plotted in order to check for constant variance (plot them against predicted values) normality (Q-Q-plots) substantial residual autocorrelation (SAC and SPAC plots)

Ljung-Box test statistic Let and define If the correct ARMA(p,q)-model is estimated, then Q *,K follows a Chi-square distribution with K – p – q degrees of freedom. Hence, excessive values of this statistic indicates that the model has been erroneously specified. The value of K should be chosen large enough to cover what can be expected to be a set of autocorrelations that are unusually high if the model is wrong.