Analysis of financial data Anders Lundquist Spring 2010.

Slides:



Advertisements
Similar presentations
Autocorrelation Functions and ARIMA Modelling
Advertisements

Regression Analysis.
Properties of Least Squares Regression Coefficients
Dates for term tests Friday, February 07 Friday, March 07
COMM 472: Quantitative Analysis of Financial Decisions
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
The Simple Regression Model
Using SAS for Time Series Data
Model Building For ARIMA time series
CHAPTER 8 MULTIPLE REGRESSION ANALYSIS: THE PROBLEM OF INFERENCE
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Time Series Building 1. Model Identification
Correlation and regression
Bivariate Regression Analysis
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 4 Multiple Regression.
KYIV SCHOOL OF ECONOMICS Financial Econometrics (2nd part): Introduction to Financial Time Series May 2011 Instructor: Maksym Obrizan Lecture notes II.
ARIMA Forecasting Lecture 7 and 8 - March 14-16, 2011
Chapter 11 Multiple Regression.
Economics 20 - Prof. Anderson
Modern methods The classical approach: MethodProsCons Time series regression Easy to implement Fairly easy to interpret Covariates may be added (normalization)
Volatility Chapter 9 Risk Management and Financial Institutions 2e, Chapter 9, Copyright © John C. Hull
K. Ensor, STAT Spring 2005 The Basics: Outline What is a time series? What is a financial time series? What is the purpose of our analysis? Classification.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Lecture 5 Correlation and Regression
Time-Varying Volatility and ARCH Models
Objectives of Multiple Regression
BOX JENKINS METHODOLOGY
AR- MA- och ARMA-.
Inference for regression - Simple linear regression
Regression Method.
Time-Series Analysis and Forecasting – Part V To read at home.
STAT 497 LECTURE NOTES 2.
Ms. Khatijahhusna Abd Rani School of Electrical System Engineering Sem II 2014/2015.
#1 EC 485: Time Series Analysis in a Nut Shell. #2 Data Preparation: 1)Plot data and examine for stationarity 2)Examine ACF for stationarity 3)If not.
Intervention models Something’s happened around t = 200.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Value at Risk Chapter 16. The Question Being Asked in VaR “What loss level is such that we are X % confident it will not be exceeded in N business days?”
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
1 Regression Analysis The contents in this chapter are from Chapters of the textbook. The cntry15.sav data will be used. The data collected 15 countries’
Generalised method of moments approach to testing the CAPM Nimesh Mistry Filipp Levin.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
1 B IVARIATE AND MULTIPLE REGRESSION Estratto dal Cap. 8 di: “Statistics for Marketing and Consumer Research”, M. Mazzocchi, ed. SAGE, LEZIONI IN.
STAT 497 LECTURE NOTE 9 DIAGNOSTIC CHECKS 1. After identifying and estimating a time series model, the goodness-of-fit of the model and validity of the.
MULTIVARIATE TIME SERIES & FORECASTING 1. 2 : autocovariance function of the individual time series.
Correlation & Regression Analysis
Chapter 8: Simple Linear Regression Yang Zhenlin.
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
1 Chapter 5 : Volatility Models Similar to linear regression analysis, many time series exhibit a non-constant variance (heteroscedasticity). In a regression.
Computational Finance II: Time Series K.Ensor. What is a time series? Anything observed sequentially (by time?) Returns, volatility, interest rates, exchange.
Analysis of financial data Anders Lundquist Spring 2010.
MODEL DIAGNOSTICS By Eni Sumarminingsih, Ssi, MM.
Analysis of financial data Anders Lundquist Spring 2010.
Analysis of Financial Data Spring 2012 Lecture 9: Volatility Priyantha Wijayatunga Department of Statistics, Umeå University
Analysis of Financial Data Spring 2012 Lecture 4: Time Series Models - 1 Priyantha Wijayatunga Department of Statistics, Umeå University
The simple linear regression model and parameter estimation
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Chapter 6: Autoregressive Integrated Moving Average (ARIMA) Models
Charles University Charles University STAKAN III
STAT 497 LECTURE NOTE 9 DIAGNOSTIC CHECKS.
Volatility Chapter 10.
Product moment correlation
BOX JENKINS (ARIMA) METHODOLOGY
Presentation transcript:

Analysis of financial data Anders Lundquist Spring 2010

ARCH-modelling Recall that EWMA and MA are not model based. They are merely the methods invented to track the changes in volatility observed in returns. Another way to tackle the problem of estimating and forecasting volatilities is to build models for the conditional variance Var (Y t | Y t-1, Y t-2,...) in the same way that we did for the conditional expectation E(Y t | Y t-1, Y t-2,...) (linear autoregressions).

The sample variance was defined as (assuming the returns have zero mean) This is an estimator of  2 = V(Y t ) = E( ). It is clear from the definition of  2 that (the squared returns) contain information about the variance. Exploring and finding models for is then equivalent to studying the volatility of the returns, in the same way that exploring and finding models for Y 1, Y 2,..., Y n is equivalent to studying the mean of the returns.

We have already earlier explored the squared returns. We have, for instance, computed s 2 and sometimes assumed that it converged (stationarity) to  2. Assuming stationarity of the variance is equivalent to assume the stationarity of the mean of the squared returns! What we have not done yet is to explore the dependence structure of In the same way as we did for returns, we can look at the autocorrelations of squared returns.

ACF / PACF of log-returns

ACF / PACF of squared log-returns

According to the partial autocorrelation plots, we conclude that the log-returns of Nordea seem to be uncorrelated while the squared log-returns look correlated. This is a fairly common behavior of financial asset log-returns!

Now, we are ready to define a model for conditional variance by using the model for linear dependence of the squared log-returns. The squared log-returns are linearly dependent. A model for this linear dependence can be specified as we have done for the plain log-returns. Var (Y t | Y t-1, Y t-1,...) = Alternatively, we may write

It is more convenient to write down a model for the log-return so that, in particular, both the conditional mean and variance may be modeled together. Thus, we write a “total” model Y t = μ t +  t  t, with The parameters  0,  1,...,  p must be positive.

This model for the returns specify both the mean and the variance of the log-returns. In general, one may have μ t = f (Y t−1,...,Y t−p ). The conditional mean may vary depending on the history of log-returns.

Also, the model considers a variance. If, then the variance is constant through time. Otherwise, the conditional variance varies depending on the history of the log-returns. If the mean of the log-returns (μ t ) can be assumed to be zero, we have an ARCH model (AutoRegressive Conditional Heteroskedasticity model):

We have seen that the log-returns of Nordea seemed uncorrelated and that their sample mean is very close to zero. The squared log-returns, on the other hand, seemed correlated which implies that an ARCH model is suitable. (model the linear dependence in the squared log-returns). To decide: How many days p should we include in the model for the variance

This question is solved with the partial autocorrelations. Partial autocorrelations of the squared log-returns are used to determinate p in the same way that partial autocorrelations of log- returns were used to determine the order of a linear autoregression. Justification: we want to find a linear autoregression model for the squared log- returns!

PACF

The parameters of an ARCH model can be estimated as the parameters of a linear autoregression, again because an ARCH model is a linear model for the squared log-returns. Least squares can be used. This can be done with SPSS. When a negative value for  i is obtained, it should be set to zero. (since ”anything” squared must be above, or possibly equal to, zero) However, for ARCH models, a better fit is obtained by assuming a distribution for  t in To use this you need specialized software (SPSS cannot do it).

From the partial autocorrelations of the squared log-returns we decide that we should use an ARCH model of order 11. We estimate the parameters using SPSS and obtain:

SPSS output

A few estimates negative (but the variables look non-significant) Alternatively, we can use only those coefficients that show significance for respective parameters. In this case, we would use lag 1, lag 4, lag 6, lag 7, lag 9, lag 10 and lag 11 plus a constant. In order to do this, we have to save all these lagged variables and perform linear regression.

SPSS output

Note: the estimated coefficients from this reduced model are somewhat different from the corresponding coefficients from the full model with 11 lags…

Forecast, squared log-returns

A usual model for the error term  t is that it is normally distributed. An ARCH model with normal error term  t will typically imply that Y t (the log-returns) are not normally distributed but instead have large kurtosis. So eventually we have a model which explains the observed behaviour of the log-returns. Consider a series of simulated log-returns, using the significant estimated parameters from the ARCH(11) as input.

Simulated observations, ARCH(11)

Comparison, observed in the left column, simulated ARCH in the right

Comparison, numerical summaries Simulated observations, ARCH(11) MeanStDevSkewnessKurtosis 0, Observed values for the log-returns

Now, the volatility of returns may be estimated using different estimators, moving averages, exponentially weighted moving averages and also by using a model for conditional volatility (ARCH). A natural question arise: How can we decide which estimator to choose? There are several ways to proceed to compare the performance of ARCH and EWMA (MA) estimators but this topic is not included in this course (see, e.g., Market Models by Carol Alexander, 2001) We just make simple comparisons by plotting different estimators.

Order of graphs on next page Observed squared log- returns ARCH MA(30) EWMA(0.94)

Summary We have seen that ARCH models give an account of excess kurtosis often present in financial data. These models succeed in capturing departures from normality in asset log-returns. Consequently, they may be useful alternatives to MA-methods