USING A GENETIC PROGRAM TO PREDICT EXCHANGE RATE VOLATILITY Christopher J. Neely Paul A. Weller.

Slides:



Advertisements
Similar presentations
Agenda of Week V. Forecasting
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
Model Assessment, Selection and Averaging
Chapter 7 – Classification and Regression Trees
CMPUT 466/551 Principal Source: CMU
Chapter 7 – Classification and Regression Trees
On Predictability and Profitability: Would AI induced Trading Rules be Sensitive to the Entropy of time Series Nicolas NAVET – INRIA France
Exponential Smoothing Methods
Validation and Monitoring Measures of Accuracy Combining Forecasts Managing the Forecasting Process Monitoring & Control.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
CAViaR : Conditional Value at Risk By Regression Quantiles Robert Engle and Simone Manganelli U.C.S.D. July 1999.
MOVING AVERAGES AND EXPONENTIAL SMOOTHING
Evaluating Hypotheses
Chapter 3 Forecasting McGraw-Hill/Irwin
Chapter 13 Forecasting.
T T18-05 Trend Adjusted Exponential Smoothing Forecast Purpose Allows the analyst to create and analyze the "Trend Adjusted Exponential Smoothing"
The generalized Additive Nonparametric GARCH Model --With application to the Chinese stock market Ai Jun Hou Department of Economics School of Economics.
Prediction and model selection
Judgment in Forecasting, Forecast Accuracy, Moving Averages and Decomposition Lecture 2 February 23, 2010.
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Volatility Chapter 9 Risk Management and Financial Institutions 2e, Chapter 9, Copyright © John C. Hull
Forecasting McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
15-1 Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Forecasting Chapter 15.
Statistical Treatment of Data Significant Figures : number of digits know with certainty + the first in doubt. Rounding off: use the same number of significant.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
Business Forecasting Chapter 5 Forecasting with Smoothing Techniques.
Relationships Among Variables
Forecasting Chapter 15.
Forecasting Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill.
1 What does genetic programming teach us about the foreign exchange market ? Chris Neely* Paul Weller † Rob Dittmar** December 1-2, 1998 *Economist, Federal.
1 MBF 2263 Portfolio Management & Security Analysis Lecture 2 Risk and Return.
Presented by Tienwei Tsai July, 2005
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
Introduction to Management Science
Chapter 9 – Classification and Regression Trees
McGraw-Hill/Irwin Copyright © 2002 by The McGraw-Hill Companies, Inc. All rights reserved. 3-2 Business Forecasting with Accompanying Excel-Based ForecastX™
N-tuple S&P Patterns Across Decades, 1950s to 2011
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
1 DSCI 3023 Forecasting Plays an important role in many industries –marketing –financial planning –production control Forecasts are not to be thought of.
MBA.782.ForecastingCAJ Demand Management Qualitative Methods of Forecasting Quantitative Methods of Forecasting Causal Relationship Forecasting Focus.
Operations Research II Course,, September Part 6: Forecasting Operations Research II Dr. Aref Rashad.
Measures of central tendency are statistics that express the most typical or average scores in a distribution These measures are: The Mode The Median.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Forecasting Chapter 9. Copyright © 2013 Pearson Education, Inc. publishing as Prentice Hall Define Forecast.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
15-1 Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Forecasting Chapter 15.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
Estimating Volatilities and Correlations Chapter 21.
15-1 Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Forecasting Chapter 15.
Run length and the Predictability of Stock Price Reversals Juan Yao Graham Partington Max Stevenson Finance Discipline, University of Sydney.
1 Chapter 5 : Volatility Models Similar to linear regression analysis, many time series exhibit a non-constant variance (heteroscedasticity). In a regression.
1 BABS 502 Moving Averages, Decomposition and Exponential Smoothing Revised March 14, 2010.
Estimating Volatilities and Correlations
FORECASTING METHODS OF NON- STATIONARY STOCHASTIC PROCESSES THAT USE EXTERNAL CRITERIA Igor V. Kononenko, Anton N. Repin National Technical University.
Geo479/579: Geostatistics Ch12. Ordinary Kriging (2)
3-1Forecasting CHAPTER 3 Forecasting McGraw-Hill/Irwin Operations Management, Eighth Edition, by William J. Stevenson Copyright © 2005 by The McGraw-Hill.
MBF1413 | Quantitative Methods Prepared by Dr Khairul Anuar 8: Time Series Analysis & Forecasting – Part 1
1 A latent information function to extend domain attributes to improve the accuracy of small-data-set forecasting Reporter : Zhao-Wei Luo Che-Jung Chang,Der-Chiang.
McGraw-Hill/Irwin Copyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 3 Forecasting.
Forecas ting Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill.
Forecast 2 Linear trend Forecast error Seasonal demand.
T T18-02 Weighted Moving Average Forecast Purpose Allows the analyst to create and analyze the "Weighted Moving Average" forecast for up to 5.
TIME SERIES MODELS. Definitions Forecast is a prediction of future events used for planning process. Time Series is the repeated observations of demand.
Volatility Chapter 10.
MBF1413 | Quantitative Methods Prepared by Dr Khairul Anuar
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Presentation transcript:

USING A GENETIC PROGRAM TO PREDICT EXCHANGE RATE VOLATILITY Christopher J. Neely Paul A. Weller

Introduction The Genetic Program Data and Implementation Experimental Designs Results Conclusion Discussion Outline

Introduction Exchange rate volatility displays considerable persistence. Large movements in prices tend to be followed by more large moves, producing positive serial correlation in absolute or squared returns. Engle (1982) ARCH model Bollerslev (1986) GARCH model This paper investigates the performance of a genetic program as a non-parametric procedure for forecasting volatility in the foreign exchange market. genetic program

Strength and Weakness Strength Genetic programs have the ability to detect patterns in the conditional mean of foreign exchange and equity returns that are not accounted for by standard statistical models (Neely, Weller, and Dittmar 1997; Neely and Weller 1999 and 2001; Neely 2000). Weakness Over fitting

The Genetic Program Function Set Data functions data, average, max, min, and lag. Four data series Three more complex data functions geo, mem, arch5 geomemarch5 An Example of a GP Tree Volatility

Function Set plus, minus, times, divide, norm, log, exponential, square root, and cumulative standard normal distribution function.

Four data series Daily Returns Integrated Volatility the sum of squared intraday returns The sum of the absolute value of intraday returns Number of days until the next business day

The function geo returns the following weighted average of 10 lags of past data. This function can be derived from the prediction of an IGARCH specification with parameter , where we constrain  to satisfy 0.01    0.99 and lags are truncated at 10.

The function mem returns a weighted sum similar to that which would be obtained from a long memory specification.  j>0 h j =1 d is determined by the genetic program and constrained to satisfy -1< d <1.

The function arch5 permits a flexible weighting of the five previous observations. h j are provided by the genetic program and constrained to lie within {-5,5} and to sum to one.

Volatility Since true volatility is not directly observed, it is necessary to use an appropriate proxy in order to assess the performance of the genetic program. Use the ex post squared daily return. Andersen and Bollerslev (1998) A better approach is to sum intraday returns to more accurately measure true daily volatility (i.e., integrated volatility).

S i,t is the i-th observation on date t  2 i,t is the measure of integrated volatility on date t. More precisely, daily volatility is calculated from 1700 GMT to 1700 GMT. Using five intraday observations represents a compromise between the increase in accuracy generated by more frequent observations and the problems of data handling and availability that arise as one moves to progressively higher frequencies of intraday observation.

Data and Implementation Dollar / German mark (DEM) Dollar / Japanese yen (JPY) June 1975 to September Training period June 1975 – December 1979 Selection period January 1980 – December 30, 1986 Out-of-sample period December 31, 1986 – September 21, 1999 The sources of the data Step

The sources of the data

Step 1. Create an initial generation of 500 randomly generated forecast functions. 2. Measure the MSE of each function over the training period and rank according to performance. 3. Select the function with the lowest MSE and calculate its MSE over the selection period. Save it as the initial best forecast function. 4. Select two functions at random, using weights attaching higher probability to more highly-ranked functions. Apply the recombination operator to create a new function, which then replaces an old function, chosen using weights attaching higher probability to less highly- ranked functions. Repeat this procedure 500 times to create a new generation of functions.

5. Measure the MSE of each function in the new generation over the training period. Take the best function in the training period and evaluate the MSE over the selection period. If it outperforms the previous best forecast, save it as the new best forecast function. 6. Stop if no new best function appears for 25 generations, or after 50 generations. Otherwise, return to stage 4.stage 4

Experimental Designs Benchmark Fitness Function Measure of Forecast Errors Forecasts Aggregation Other Designs

Benchmark GARCH (1,1)

Fitness Function MSE MSE + Penalty

Overfitting Penalty Function for Node Complexity This consisted of subtracting an amount (0.002 * number of nodes) from the negative MSE. This modification is intended to bias the search toward functions with fewer nodes, which are simpler and therefore less prone to overfit the data.

Measure of Forecast Errors mean square error mean absolute error R-square mean forecast bias kernel estimates of the error densities.

Forecasts Aggregation The forecasts were aggregated in one of two ways. Mean: The equally-weighted forecast is the arithmetic average of the forecasts from each of the ten trials. Median: The median-weighted forecast takes the median forecast from the set of ten forecasts at each date.

Other Designs forecast horizon: 1, 5, 10 the number of data functions: 5, 8 penalty for complexity: absent, present

Results An example of a one-day ahead forecasting functions for the DEM An example In-sample comparison of GP and GARCH In-sample comparison Out-of- sample comparison of GP and GARCH Out-of- sample comparison Out-of- sample results using the data functions geo, mem, arch5 Out-of- sample results using the data functions geo, mem, arch5 Kernel estimates of the densities of out-of sample forecast errors Kernel estimates of the densities of out-of sample forecast errors Tests for mean forecast bias -- Newey-West correction for serial correlation Tests for mean forecast bias Summary

An example

In-sample comparisonIn-sample comparison of GP and GARCH The equally-weighted forecast is the arithmetic average of the forecasts from each of the ten trials. The median-weighted forecast takes the median forecast from the set of ten forecasts at each date.

In-Sample That is, its best relative performance is at the twenty-day horizon. The median weighted forecast is generally somewhat inferior to the equally weighted forecast.

Out-of- sample comparisonOut-of- sample comparison of GP and GARCH << <<

Out-of-Sample With MSE as the performance criterion, neither the genetic program nor the GARCH model is clearly superior. The GARCH model achieves higher R^2 in each case. But the MAE criterion clearly prefers the genetic programming forecasts.

Out-of- sample results using the data functions geo, mem, arch5using the data functions geo, mem, arch5

Effect of Advanced Functions We have established that neither imposing a penalty for complexity nor expanding the set of data functions leads to any appreciable improvement in the performance of the genetic program.a penalty for complexity

Effect of the Penalty Function This had very little effect and if anything led to a slight deterioration in out-of- sample performance.

Kernel estimates of the densities of out-of sample forecast errorsout-of sample forecast errors The appearance of greater bias in the GARCH forecasts is illusory.

The most striking feature to emerge from these figures is the apparent bias in the GARCH forecasts when compared to their genetic program counterparts.

Tests for mean forecast bias Though both forecasts are biased in the mean, the magnitude of the bias is considerably greater for the genetic program.

Summary While the genetic programming rules did not usually match the GARCH(1,1) model's MSE or R^2 at 1-and 5-day horizons, its performance on those measures was generally close. But the genetic program did consistently outperform the GARCH model on mean absolute error (MAE) and modal error bias at all horizons and on most measures at 20-day horizons.

Conclusion GP did reasonably well in forecasting out-of- sample volatility. While the GP rules did not usually match the GARCH(1,1) model’s MSE or R 2 at 1- and 5- day horizons, its performance on those measures was generally close. GP did consistently outperform the GARCH model on mean absolute error (MAE) and modal error bias at all horizons and on most measures at 20-day horizons.

Discussion Choice of Function Sets Simple Functions (Primitive Functions) Complex Functions Use of Data Functions True Volatility The Selection Period

Use of Data Functions The data functions can operate on any of the four data series we permit as inputs to the genetic program. data, average, max, min, and lag.

True Volatility The functions generated by the genetic program produce forecasts of volatility. Since true volatility is not directly observed, it is necessary to use an appropriate proxy in order to assess the performance of the genetic program.

The Selection Period What is the difference between Neely’s GP and the regular GP? Neely introduced a new termination criterion, which is based on the recent progress. This idea itself is not new. What makes Neely’s idea unique is that the progress is measured by a ``testing sample’’, which he called it the selection period.

The Selection Period Neely’s GP can be considered as another approach to avoid over-fitting. Because one characteristic of over fitting is the feature that the in-sample performance is improving, while the post-sample performance is stagnated or get worse. Use Szpiro (2001)’s three-stage development of GP in data mining as a reference. This is similar to the early stopping criterion frequently used in the artificial neural nets.