Pretests for genetic-programming evolved trading programs : “zero-intelligence” strategies and lottery trading Nicolas NAVET INRIA - AIECON NCCU

Slides:



Advertisements
Similar presentations
1 Entropy Rate and Profitability of Technical Analysis: Experiments on the NYSE US 100 Stocks Nicolas NAVET – INRIA France Shu-Heng CHEN.
Advertisements

Genetic Programming for Financial Trading Nicolas NAVET INRIA, France AIECON NCCU, Taiwan Tutorial at.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 9 Hypothesis Testing Developing Null and Alternative Hypotheses Developing Null and.
 3M is expected to pay paid dividends of $1.92 per share in the coming year.  You expect the stock price to be $85 per share at the end of the year.
On Predictability and Profitability: Would AI induced Trading Rules be Sensitive to the Entropy of time Series Nicolas NAVET – INRIA France
Genetic Algorithms as a Tool for General Optimization Angel Kuri 2001.
Data Mining Techniques Outline
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Artificial Intelligence
Simple Linear Regression
Evaluating Hypotheses
Genetic Programming and the Predictive Power of Internet Message Traffic James D Thomas Katia Sycara.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
Chapter 11 Multiple Regression.
Lecture 23 Multiple Regression (Sections )
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
16 November, 2005 Statistics in HEP, Manchester 1.
Simple Linear Regression Analysis
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Query Processing Presented by Aung S. Win.
Genetic Algorithms Overview Genetic Algorithms: a gentle introduction –What are GAs –How do they work/ Why? –Critical issues Use in Data Mining –GAs.
Agent-based Simulation of Financial Markets Ilker Ersoy.
Chapter 13: Inference in Regression
Correlation and Linear Regression
Confidence Intervals and Hypothesis Testing - II
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Fundamentals of Hypothesis Testing: One-Sample Tests
Hypothesis Testing in Linear Regression Analysis
Chapter 1: Introduction to Statistics
Genetic Programming on Program Traces as an Inference Engine for Probabilistic Languages Vita Batishcheva, Alexey Potapov
1 What does genetic programming teach us about the foreign exchange market ? Chris Neely* Paul Weller † Rob Dittmar** December 1-2, 1998 *Economist, Federal.
Presented by Bruce Vanstone for the Australian Shareholders Association.
Chap 20-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 20 Sampling: Additional Topics in Sampling Statistics for Business.
Cristian Urs and Ben Riveira. Introduction The article we chose focuses on improving the performance of Genetic Algorithms by: Use of predictive models.
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Evolution Strategies Evolutionary Programming Genetic Programming Michael J. Watts
Improved Gene Expression Programming to Solve the Inverse Problem for Ordinary Differential Equations Kangshun Li Professor, Ph.D Professor, Ph.D College.
Study on Genetic Network Programming (GNP) with Learning and Evolution Hirasawa laboratory, Artificial Intelligence section Information architecture field.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
Design of an Evolutionary Algorithm M&F, ch. 7 why I like this textbook and what I don’t like about it!
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
GENETIC ALGORITHM A biologically inspired model of intelligence and the principles of biological evolution are applied to find solutions to difficult problems.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests Statistics.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Correlation & Regression Analysis
Chapter 8: Simple Linear Regression Yang Zhenlin.
Market Efficiency.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Unit 5: Hypothesis Testing.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Stat 31, Section 1, Last Time Distribution of Sample Means –Expected Value  same –Variance  less, Law of Averages, I –Dist’n  Normal, Law of Averages,
© 2010 Pearson Prentice Hall. All rights reserved Chapter Hypothesis Tests Regarding a Parameter 10.
Genetic Algorithms An Evolutionary Approach to Problem Solving.
Evolution Strategies Evolutionary Programming
Unit 5: Hypothesis Testing
Rule Induction for Classification Using
Presented by: Dr Beatriz de la Iglesia
C.-S. Shieh, EC, KUAS, Taiwan
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Significance Tests: The Basics
Significance Tests: The Basics
Statistical Thinking and Applications
Presentation transcript:

Pretests for genetic-programming evolved trading programs : “zero-intelligence” strategies and lottery trading Nicolas NAVET INRIA - AIECON NCCU joint work with Shu-Heng Chen AIECON NCCU ICONIP 2006 – Hong Kong – October 4, 2006

2 Conclusions are - most often - inconclusive regarding the efficiency of GP Is it because market is efficient ? (i.e. nothing to learn on the training data)   Further efforts are meaningless ! Or learning algorithm is inefficient ?   Consider another ML algorithm / improvement of the GP scheme: fitness function, search intensity, high level function sets, overfitting avoidance, better genetic operators, data division scheme, etc, etc, … Pretesting may provide some evidence ! If results are not very convincing :

3 Genetic programming : a recap Generate a population of random programs Evaluate their quality (“fitness”) Create better programs by applying genetic operators, eg - mutation - combination (“crossover”)  GP is the process of evolving a population of computer programs, that are candidate solutions, according to the evolutionary principles (e.g. survival of the fittest) Solution

4 In GP, programs are represented by trees (1/3) Trees are a very general representation form :  E.g. of a trading rule : buy if expression is “true” functions terminals

5 GP for financial trading Predicting price evolution (not discussed here) Inducing technical trading rules : Training interval Validation interval Out-of-sample interval 1 ) Creation of the trading rules using GP 2) Selection of the best resulting strategies Further selection on unseen data - One strategy is chosen for out-of-sample Performance evaluation

6 Why GP is an appealing technique for financial trading ? Easy to implement / robust evolutionary technique Trading rules (TR) should adapt to a changing environment – GP may simulate this evolution Solutions are produced under a symbolic form that can be understood and analyzed GP may serve as a knowledge discovery tool (e.g. evolution of the market)

7 But one may cast doubts on GP efficiency.. Highly heuristic – no theory ! Problems on which GP has been shown not to be significantly better than random search Few clear-cut successes reported in the financial literature GP embeds little domain specific knowledge yet.. Doubts on the efficiency of GP to use the available computing time : code bloat bad at finding numerical constants best solutions are sometimes found very early in the run.. Variability of the results ! e.g. returns: , , , , , , , , , , , , , , , , , , , , ….

8 Possible pretest : measure of predictability of the financial time-series Possible pretest : measure of predictability of the financial time-series   Serial correlation   Kolmogorov complexity   Lyapunov exponent   Unit root analysis   Comparison with results on surrogate data : “shuffled” series (e.g. Kaboudan statistics)  ... Actual question : how predictable for a given horizon with a given cost function?

9 In practice, some predictability does not imply profitability.. Volatility may not be sufficient to cover round-trip transactions costs! Not the right trading instrument at hand.. typically short selling not available Prediction horizon must be large enough!

Part 2 : Pretests for GP evolved trading strategies

11 Pretest methodology Compare GP with several variants of Random search algorithms   “Zero-Intelligence Strategies” - ZIS Random trading behaviors   “Lottery trading” - LT Statistical hypotheses testing Null : GP does not outperform ZIS Null : GP does not outperform LT Issue : how to best constraint randomness ?

Pretest 1 : GP versus Zero-Intelligence strategies (=“Equivalent search intensity” Random Search (ERS) with validation stage) -Null hypothesis H : GP does not outperform equivalent random search - Alternative hypothesis is H -Null hypothesis H 1,0 : GP does not outperform equivalent random search - Alternative hypothesis is H 1,1

13 Pretest 1 : GP vs zero-intelligence strategies H 1,0 cannot be rejected – interpretation :   There is nothing to learn or GP is not very effective Training interval Validation interval Out-of-sample interval 1 ) Creation of the trading rules using GP 2) Selection of the best resulting strategies Further selection on unseen data - One strategy is chosen for out-of-sample Performance evaluation ERS

14 Pretest 4 : GP vs lottery trading Lottery trading (LT) = random trading behavior according the outcome of a r.v. (e.g. Bernoulli law) Issue 1 : if LT tends to hold positions (short, long) for less time that GP, transactions costs may advantage GP.. Issue 2 : it might be an advantage or an disadvantage for LT to trade much less or much more than GP.   ex: downward oriented market with no short-sell

15 Frequency and intensity of a trading strategy Frequency : average number of transactions per unit of time Intensity : proportion of time where a position is held For pretest 4 :   We impose that average frequency and intensity of LT is equal to the ones of GP   Implementation : generate random trading sequences having the right characteristics 0,0,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,1,1,0,1,0,0,0,0,0,0,1,1,1,1,1,1,…

16 Training interval Validation interval Out-of-sample interval 1 ) Creation of the trading rules using GP 2) Selection of the best resulting strategies Further selection on unseen data - One strategy is chosen for out-of-sample Performance evaluation Pretest 4 : implementation 0,0,1,1,1,0,0,0,0,0,1,… Lottery trading

Answering question 1 : is there anything to learn on the training data at hand ?

18 Question 1 : pretests involved Starting point: if a set of search algorithms do not outperform LT, it gives evidence that there is nothing to learn.. Pretest 4 : GP vs Lottery Trading   Null hypothesis H 4,0 : GP does not outperform LT Pretest 5 : Equivalent Random Search (ZIS) vs Lottery Trading   Null hypothesis H 5,0 : ERS does not outperform LT

19 Question 1 : some answers...  R means that the null hypothesis Hcannot be rejected – R means we should favor H  R means that the null hypothesis H i,0 cannot be rejected – R means we should favor H i,1 H 4,0 H 5,0 Interpretation Case 1 RR RR Case 2RR Case 3R RR Case 4 RR R there is nothing to learn there is something to learn there may be something to learn - ERS might not be powerful enough there may be something to learn – GP evolution process is detrimental

Answering question 2 : is GP effective ?

21 Question 2 : some answers... Question 2 cannot be answered if there is nothing to learn (case 1) Case 4 provides us with a negative answer.. In case 2 and 3, run pretest 1 : GP vs Equivalent random search Null hypothesis H 1,0 : GP does not outperform ERS If one cannot reject H 1,0 GP shows no evidence of efficiency…

Pretests at work Methodology : Draw conclusions from pretests using our own programs and compare with results in the literature [ChKuHo06] on the same time series

23 Setup : GP control parameters - same as in [ChKuHo06]

24 Setup : statistics, data, trading scheme Hypothesis testing with student t-test with a 95% confidence level Hypothesis testing with student t-test with a 95% confidence level Pretests with samples made of 50 GP runs, 50 ERS runs and 100 LT runs Pretests with samples made of 50 GP runs, 50 ERS runs and 100 LT runs Data : indexes of 3 stock exchanges Canada, Taiwan and Japan Data : indexes of 3 stock exchanges Canada, Taiwan and Japan Daily trading with short selling Daily trading with short selling Training of 3 years – Validation of 2 years Training of 3 years – Validation of 2 years Out-of-sample periods: , , Out-of-sample periods: , , Data normalized with a 250 days moving average Data normalized with a 250 days moving average

25 Results on actual data (1/2)   Evidence that there is something to learn : 4 markets out of 9 (C3,J2,T1,T3)   Experiments in [ChKuHo06], with another GP implementation, show that GP performs very well on these 4 markets   Evidence that there is nothing to learn : 3 (C1,J3,T2)   In [ChKuHo06], there is only one (C1) where GP has positive return (but less than B&H)

26 Results on actual data (2/2) GP effective : 3 markets out of 6   In these 3 markets, GP outperforms Buy and Hold – same outcome as in [ChKuHo06] Preliminary conclusion : one can rely on pretests..   When there is nothing to learn, no GP implementation did good (except in one case)   When there is something to learn, at least one implementation did good (always)   When our GP is effective, GP in [ChKuHo06] is effective too (always)

27 Further conclusion Our GP implementation is is more efficient than random search : no case where ERS outperform LT and GP did not 2. 2.But only slightly more efficient … one would expect much more cases where GP does better than LT and not ERS Our GP is actually able to take advantage of regularities in data … but only of “simple” ones Ongoing work : study the correlation between predictability measures and GP performances

28 References [ChKuHo06] S.-H. Chen and T.-W. Kuo and K.-M. Hoi, “Genetic Programming and Financial Trading: How Much about What We Know”. Handbook of Financial Engineering, Kluwers, [ChKuHo06] S.-H. Chen and T.-W. Kuo and K.-M. Hoi, “Genetic Programming and Financial Trading: How Much about What We Know”. Handbook of Financial Engineering, Kluwers, [Kab00] M. Kaboudan, “Genetic Programming Prediction of Stock Prices”, Computational Economics, vol16, [Kab00] M. Kaboudan, “Genetic Programming Prediction of Stock Prices”, Computational Economics, vol16, 2000.

29 ?

30 Conclusions are - most often - inconclusive regarding the efficiency of GP “Annual returns with the GP induced technical trading rules is x%” If negative, market is efficient or GP ineffective ?? If positive, mere luck or GP is effective ?? Good/bad wrt to other search techniques ?? Worth to further improve/ optimize GP ?? Pretests provide some evidence whether 1.There is something to be learned from the data 2.GP is effective at this task

31 Equivalent search intensity Starting point : 2 search algorithms have similar search intensity if they create the same number of solutions over the course of their execution Problem : same solutions tends to be rediscovered over time and are not re-evaluated – rate of discovery strongly depends on the search technique / implementation Refined definition : similar search intensity if same number of “truly” different solutions – here truly means syntactically different

32 Other zero-intelligence (but less meaningful) strategies one can think of Pretest 2 : GP versus equivalent random search without validation   May give some insight into effectiveness of validation to fight overfitting.. but little overfitting with random search thus usefulness is dubious.. Pretest 3 : GP versus equivalent random search without training and validation : random trees applied directly out-of-sample   Bias in randomness induced by the GP language..

33 Pretest 1 : GP vs zero-intelligence strategies Implementation : Execute multiple GP runs – record average number of syntactically different individuals Random search is implemented with the initial population of GP – adjust size of population to obtain “equivalent search intensity” H 1,0 cannot be rejected – interpretation :   There is nothing to learn or GP is not effective H 1,1 should be rejected – interpretation :   There may be something to learn and GP may be effective..

34 Empty slide XXX