LCS Case Studies BRANDEN PAPALIA, JAMES PATRICK, MICHAEL STEWART FACULTY OF ENGINEERING, COMPUTING AND MATHEMATICS.

Slides:



Advertisements
Similar presentations
A real-time adaptive trading system using Genetic Programming QF5205 : Topics in Quantitative Finance Donny Lee Nadim Mouchonnet.
Advertisements

Decision Tree Approach in Data Mining
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
ABSTRACT Technical Analysis is the study of historical price trends in order to predict future price movements. Traders use various forms of Technical.
SUPPORT VECTOR MACHINES PRESENTED BY MUTHAPPA. Introduction Support Vector Machines(SVMs) are supervised learning models with associated learning algorithms.
TEMPLATE DESIGN © Genetic Algorithm and Poker Rule Induction Wendy Wenjie Xu Supervised by Professor David Aldous, UC.
Genetic algorithms applied to multi-class prediction for the analysis of gene expressions data C.H. Ooi & Patrick Tan Presentation by Tim Hamilton.
Data Mining Techniques Outline
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
1. Elements of the Genetic Algorithm  Genome: A finite dynamical system model as a set of d polynomials over  2 (finite field of 2 elements)  Fitness.
Object Recognition Using Genetic Algorithms CS773C Advanced Machine Intelligence Applications Spring 2008: Object Recognition.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
Feature Selection for Regression Problems
Supervised classification performance (prediction) assessment Dr. Huiru Zheng Dr. Franscisco Azuaje School of Computing and Mathematics Faculty of Engineering.
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
1 Abstract This study presents an analysis of two modified fuzzy ARTMAP neural networks. The modifications are first introduced mathematically. Then, the.
Ensemble Learning: An Introduction
Evaluating Hypotheses
Genetic Programming and the Predictive Power of Internet Message Traffic James D Thomas Katia Sycara.
Basic concepts of Data Mining, Clustering and Genetic Algorithms Tsai-Yang Jea Department of Computer Science and Engineering SUNY at Buffalo.
Genetic Algorithm Genetic Algorithms (GA) apply an evolutionary approach to inductive learning. GA has been successfully applied to problems that are difficult.
7/2/2015Intelligent Systems and Soft Computing1 Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Introduction,
Genetic Algorithm What is a genetic algorithm? “Genetic Algorithms are defined as global optimization procedures that use an analogy of genetic evolution.
Chapter 5 Data mining : A Closer Look.
Learning Classifier Systems Dominic Cockman, Jesper Madsen, Qiuzhen Zhu 1.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Genetic Algorithm.
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
Comparing the Parallel Automatic Composition of Inductive Applications with Stacking Methods Hidenao Abe & Takahira Yamaguchi Shizuoka University, JAPAN.
Evolutionary Intelligence
© Negnevitsky, Pearson Education, CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University.
Efficient Model Selection for Support Vector Machines
1 Learning Classifier Systems Andrew Cannon Angeline Honggowarsito 1.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Working with Spreadsheets S S T : S P R E A D S H E E T S SST 2 Objectives 1.Perform data entry tasks 2.Use formulae and functions in worksheet calculations.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Chapter 8 The k-Means Algorithm and Genetic Algorithm.
Benk Erika Kelemen Zsolt
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
GENETIC ALGORITHMS.  Genetic algorithms are a form of local search that use methods based on evolution to make small changes to a popula- tion of chromosomes.
© Negnevitsky, Pearson Education, Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Introduction,
Wireless communications and mobile computing conference, p.p , July 2011.
1 Appendix D: Application of Genetic Algorithm in Classification Duong Tuan Anh 5/2014.
Learning Classifier Systems BRANDEN PAPALIA, MICHAEL STEWART, JAMES PATRICK FACULTY OF ENGINEERING, COMPUTING AND MATHEMATICS.
Genetic Algorithms Genetic algorithms provide an approach to learning that is based loosely on simulated evolution. Hypotheses are often described by bit.
Franciszek Seredynski, Damian Kurdej Polish Academy of Sciences and Polish-Japanese Institute of Information Technology APPLYING LEARNING CLASSIFIER SYSTEMS.
Learning Classifier Systems (Introduction) Muhammad Iqbal Evolutionary Computation Research Group School of Engineering and Computer Science Victoria University.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Genetic Algorithms (in 1 Slide) l GA: based on an analogy to biological evolution l Each.
GENETIC ALGORITHM Basic Algorithm begin set time t = 0;
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
UCSpv: Principled Voting in UCS Rule Populations Gavin Brown, Tim Kovacs, James Marshall.
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Ensemble Classifiers.
Big data classification using neural network
Evolving Decision Rules (EDR)
Comparing Genetic Algorithm and Guided Local Search Methods
The
iSRD Spam Review Detection with Imbalanced Data Distributions
Department of Electrical Engineering
EE368 Soft Computing Genetic Algorithms.
Presentation transcript:

LCS Case Studies BRANDEN PAPALIA, JAMES PATRICK, MICHAEL STEWART FACULTY OF ENGINEERING, COMPUTING AND MATHEMATICS

The University of Western Australia Mining Imbalanced Data

The University of Western Australia XCS Overview  XCS evolves a population of classifiers P  Each classifier is a rule where condition -> action  Rule also contains a set of parameters: Prediction pthe payoff prediction Prediction error eestimate the error between p and payoff Fitness Fevaluates the classifier’s accuracy Numerosity num the number of copies of this classifier in the population

The University of Western Australia Procedure  At each iteration the XCS builds the match set M  M contains all classifiers in P that match the input  Two types of selection regimes may be applied: Pure-explore regime selects action with highest prediction p Pure-exploit regime selects action with highest fitness F  Pure-explore is typically used during training, whereas pure-exploit is used when the system predicts new unknown instances  The final selected classifiers form the action set A

The University of Western Australia  The chosen action from A is sent to the environment  Environment returns a reward R that the XCS uses to update the parameters of the classifiers in the population  The prediction p, error e, and fitness F is updated in each iteration Updating Parameters

The University of Western Australia 1. Prediction  The prediction is adjusted using the formula: p = p + β( R – p) Where: p is the prediction β is the learning rate (0 < β < 1) R is the reward 2. Error  The prediction is adjusted using the formula: e = e + β(|R – p| - e) Where: e is the error β is the learning rate (0 < β < 1) R is the reward p is the prediction

The University of Western Australia 3. Accuracy  XCS first computes accuracy k: Where: Ɛ 0 is the maximum error that the classifier can take to be considered accurate a and v control the rate and decline in accuracy 4. Fitness  The accuracy k is used to determine the relative accuracy k’ of the classifier in A  Fitness is updated using k’: F = F + β( k’ – F)

The University of Western Australia Discovery Component  A genetic algorithm (GA) is applied to the action set  The GA selects two parents from A, and copies them to create two new classifiers  These new classifiers are introduced into the population

The University of Western Australia Multiplexer  Orriols-Puig & Ester Bernado-Mansilla used an imbalanced multiplexer to investigate how class imbalances affect XCS  Multiplexer: Binary string of size L First log 2 L bits are address bits Remaining bits are position bits Output is the value of the position bit indicated by the decimal value of the address bits

The University of Western Australia XCS on Imbalanced Multiplexer  This case study ran an XCS on an 11-bit multiplexer  This imbalanced multiplexer deliberately undersamples the class labeled as ‘1’.  Imbalance ratio ir refers to the ratio between the number of instances sampled of the majority class TN and minority class TP  Correct classification of the majority class (TN rate) quickly raised to 100% for any imbalance ratio tested  Correct classification of the minority class (TP rate) is only 100% for ir < 16  After ir > 64, the XCS classifies all inputs as if they were in the majority class

The University of Western Australia Modelling Parameter’s Bounds  Orriols-Puig & Ester Bernado-Mansilla sought to derive a bound on ir, whereby the XCS will consider overgeneral classifiers as accurate  Overgeneral classifiers have a high error in inbalanced datasets, as they will often incorrectly classify each class  However, as ir increases, the proportion of minority class decreases, causing overgeneral classifiers to have a lower error  The case study finds that the bound on ir scales proportionally to the maximum reward, and inversely proportional to the max error to be considered accurate

The University of Western Australia Online Configuration of XCS  Case study uses an online adaption algorithm to automatically tune the XCS parameters

The University of Western Australia β Adaption  β (learning rate) adapted in order to align the prediction value of the overgeneral classifier to the theoretical prediction  Consider worst case (classifier receives 1 minority class instance and ir n instances of the majority class) and compute the error  Calculates a new learning rate based on this error

The University of Western Australia Using LCS to Mine Imbalanced Datasets  Case study looked at 25 real-world two-class problems, each with different imbalance ratios and characteristics

The University of Western Australia Results  Results show that XCS outperformed other methods in twelve occasions

The University of Western Australia Results (contd)

The University of Western Australia Summary  This case study showed that XCS can solve the multiplexer problem for ir < 32  Also presented an algorithm that modified the parameters of an XCS during runtime to provide a significant improvement to classification in imbalanced datasets  XCS is comparable with other learning algorithms in the realm of imbalanced data mining

The University of Western Australia Foreign Exchange Trading

The University of Western Australia Foreign Exchange Trading  Price prediction Technical (Not featured in our case study) Historical  Tracding method Long position Short position

The University of Western Australia The Data Set  Six sets of financial time series  5 Currencies United States Dollar (USD) German Deutsche Mark (DEM) British Pound (GBP) Japanese Yen (JPY) Swiss Franc (CHF)  A set contains 3 daily data elements and span Jan 1,1974 to Oct 11, 1995  The three elements: Daily interest rates for the base currency Daily exchange rate Interest rates for the counter currencies

The University of Western Australia Foreign exchange Trading – LC System introduction  Implementation set out to prove LCS’ competitiveness versus Genetic Programing  Utilises NewBoole, a LCS technique similar to Zeroth level Classifier System (ZCS)  NewBoole learning and reward strategy Supervised online learning Symmetrical penalty-payoff strategy  NewBoole Key parameters N – Classifier Population Size Ρ – Genetic Algorithm request rate µ – Mutation rate x – crossover rate m – mutation range β – learning rate

The University of Western Australia Learning Classifier Implementation  Time based trading indicators  Each classifier can either advocate a long or a short action for a currency for each time, t  At time t+1 the excess return (the interest that could have been earnt it no trade was made) that could have been made is compared with the chosen outcome Excess for long position Excess for a short position

The University of Western Australia Learning Classifier implementation continued  After excess return is calculated, transaction costs are subtracted  After the comparison is made the correctness of the position is established  Classifiers are categorised based on the profit they would have generated for the single transaction instance being considered  Rewards are distributed Transaction cost

The University of Western Australia Results Learning classifier system performance (Jan 1, 1981 to Sep 30, 1995) Genetic Programming (Jan 1, 1981 to Sep 30, 1995)

The University of Western Australia Notable Observations  LCS results, while not as good as GP results, did not require offline learning to be achieved  Lower rule evaluations under LCS (1,000 Million to 26 Million)  LCS can suffer quick fitness delay following incorrect trades  LCS had a tendency to advocate the same trade after many trials

The University of Western Australia References Albert Orriols-Puig, E. B.-M. (2008). Mining Imbalanced Data with Learning Classifier Systems. In L. Bull, E. Bernado-Mansilla, & J. Holmes, Learning Classifier Systems in Data Mining (pp ). Springer-Verlag Berlin Heidelberg. Chawla, N. V. (2005). Data Mining For Imbalanced Datasets: An Overview. In O. Maimon, & L. Rokach, The Data Mining and Knowledge Discovery Handbook (pp ). Springer US. Stone, C., & Bull, L. (2008). Foreign Exchange Trading Using a Learning Classifier System. In L. Bull, J. Holmes, & E. Bernado-Mansilla, Learning Classifier Systems in Data Mining (pp ). Springer Berlin Heidelberg.