1 Appendix D: Application of Genetic Algorithm in Classification Duong Tuan Anh 5/2014.

Slides:



Advertisements
Similar presentations
1 CS 391L: Machine Learning: Rule Learning Raymond J. Mooney University of Texas at Austin.
Advertisements

Rule Generation from Decision Tree Decision tree classifiers are popular method of classification due to it is easy understanding However, decision tree.
Classification Algorithms
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Introduction to Data Mining by Tan, Steinbach,
Data Mining Classification This lecture node is modified based on Lecture Notes for Chapter 4/5 of Introduction to Data Mining by Tan, Steinbach, Kumar,
Lecture Notes for Chapter 4 (2) Introduction to Data Mining
ICS320-Foundations of Adaptive and Learning Systems
Data Mining Tri Nguyen. Agenda Data Mining As Part of KDD Decision Tree Association Rules Clustering Amazon Data Mining Examples.
Classification Techniques: Decision Tree Learning
A survey of Evolutionary Algorithms for Data Mining and Knowledge Discovery Alex Freitas (2001) A look at Genetic Algorithms(GA) and Genetic Programming(GP)
Lecture Notes for Chapter 4 Part III Introduction to Data Mining
Mining Association Rules. Association rules Association rules… –… can predict any attribute and combinations of attributes … are not intended to be used.
Decision Tree Algorithm
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
Basic Data Mining Techniques Chapter Decision Trees.
Basic Data Mining Techniques
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
ROC Curves.
Covering Algorithms. Trees vs. rules From trees to rules. Easy: converting a tree into a set of rules –One rule for each leaf: –Antecedent contains a.
Basic concepts of Data Mining, Clustering and Genetic Algorithms Tsai-Yang Jea Department of Computer Science and Engineering SUNY at Buffalo.
Classification II.
Data Mining: A Closer Look Chapter Data Mining Strategies (p35) Moh!
Classification.
MACHINE LEARNING. What is learning? A computer program learns if it improves its performance at some task through experience (T. Mitchell, 1997) A computer.
Gini Index (IBM IntelligentMiner)
Chapter 7 Decision Tree.
Repository Method to suit different investment strategies Alma Lilia Garcia & Edward Tsang.
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Machine Learning Chapter 3. Decision Tree Learning
Inductive learning Simplest form: learn a function from examples
Mohammad Ali Keyvanrad
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Lecture 8: 24/5/1435 Genetic Algorithms Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Basic Data Mining Technique
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
Machine Learning Overview
Feature Selection: Why?
Categorical data. Decision Tree Classification Which feature to split on? Try to classify as many as possible with each split (This is a good split)
Informed search algorithms Chapter 4. Outline Best-first search Greedy best-first search A * search Heuristics.
1 Learning Chapter 18 and Parts of Chapter 20 AI systems are complex and may have many parameters. It is impractical and often impossible to encode all.
Classification Techniques: Bayesian Classification
Decision Trees. MS Algorithms Decision Trees The basic idea –creating a series of splits, also called nodes, in the tree. The algorithm adds a node to.
CS690L Data Mining: Classification
Chapter 20 Data Analysis and Mining. 2 n Decision Support Systems  Obtain high-level information out of detailed information stored in (DB) transaction-processing.
Decision Tree (Rule Induction)
Practical Issues of Classification Underfitting and Overfitting –Training errors –Generalization (test) errors Missing Values Costs of Classification.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Machine Learning A Quick look Sources: Artificial Intelligence – Russell & Norvig Artifical Intelligence - Luger By: Héctor Muñoz-Avila.
Chapter 6. Classification and Prediction Classification by decision tree induction Bayesian classification Rule-based classification Classification by.
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Evaluating Classification Performance
Classification & Prediction — Continue—. Overfitting in decision trees Small training set, noise, missing values Error rate decreases as training set.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Genetic Algorithms (in 1 Slide) l GA: based on an analogy to biological evolution l Each.
1 Classification: predicts categorical class labels (discrete or nominal) classifies data (constructs a model) based on the training set and the values.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Decision Tree. Classification Databases are rich with hidden information that can be used for making intelligent decisions. Classification is a form of.
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
Genetic Programming. What is Genetic Programming? GP for Symbolic Regression Other Representations for GP Example of GP for Knowledge Discovery Outline.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
University of Essex Alma Lilia García Almanza Edward P.K. Tsang Simplifying Decision Trees Learned by Genetic Programming.
Evolving Decision Rules (EDR)
Presented by: Dr Beatriz de la Iglesia
An evolutionary approach to solving complex problems
Features & Decision regions
Introduction to Data Mining, 2nd Edition by
Decision Tree Concept of Decision Tree
©Jiawei Han and Micheline Kamber
Decision Tree (Rule Induction)
COSC 4368 Intro Supervised Learning Organization
Presentation transcript:

1 Appendix D: Application of Genetic Algorithm in Classification Duong Tuan Anh 5/2014

2 Classification with Decision trees Class No Yes No Yes No Yes No Training data

3 Decision tree There exists the algorithm to create a decision tree from the training set (ID3, C4.5)

4 Classification rules from decision tree Represent the knowledge in the form of IF-THEN rules One rule is created for each path from the root to a leaf Each attribute-value pair along a path forms a conjunction The leaf node holds the class prediction Rules are easier for humans to understand. Example IF age = “<=30” AND student = “no” THEN buys_computer = “no” IF age = “<=30” AND student = “yes” THEN buys_computer = “yes” IF age = “31…40” THEN buys_computer = “yes” IF age = “>40” AND credit_rating = “excellent” THEN buys_computer = “no” IF age = “>40” AND credit_rating = “fair” THEN buys_computer = “yes”

5 GA for classification rule discovery Individual representation  Each individual encodes a single classification rule  Each rule is represented as a bit string Example: Instances in the training set are describe by two Boolean attributes A 1 and A 2 and two classes: C 1 and C 2  Rule: IF A 1 AND NOT A 2 THEN C 2  bit string “100”  Rule: IF NOT A 2 AND NOT A 2 THEN C 1  bit string “001”  If the attribute has k values, k > 2 then k bits are used to encode the attribute values. Classes can be encoded in a similar fashion.

6 Genetic operators for rule discovery Generalizing/Specializing Crossover  Overfitting: a situation in which a rule is covering one training example.  generalization  Underfitting: a situation in which a rule is covering too many training examples.  specialization  The generalizing/specialization crossover operators can be implemented as the logical OR and AND, respectively.  Example: Two crossover points children produced by children produced by Parents generalization crossover specialization crossover 0 | 1 0 | 1 0 | 1 1 | 1 0 | 0 0 | 1 1 | 0 1 | 0 1 | 1 1 | 0 1 | 0 0 | 0 OR AND

7 Fitness function Let a rule be of the form: IF A THEN C where A is the antecedent and C is the predicted class. Predictive accuracy of a rule called confidence factor (CF) is defined: CF = |A  C|/|A| |A|: the number of examples satisfying all the conditions in the antecedent A |A  C|: the number of examples that both satisfy the antecedent A and have the class predicted by the consequent C. Example: A rule covers 10 examples (i.e. |A| = 10), in which 8 examples have the class predicted by the rule (i.e. |A & C| = 8), then CF of the rule is CF = 80%. The performance of a rule can be summarized by a matrix called a confusion matrix.

8 TP = True positives = Number of examples satisfying A and C FP = False positives = Number of examples satisfying A but not C FN = False negatives = Number of examples not satisfying A but satisfying C TN = True negatives = Number of examples not satisfying A nor C CF measure is defined in terms of the above notation: CF = TP/(TP + FP). Confusion matrix

9 Fitness function (cont.) We can know measure the predictive measure of a rule by taking into account not only its CF but also a measure of how “complete” a rule is. Completeness of the rule: what is the proportion of examples having the predicted class C that is actually covered by the rule antecedent. The rule completeness measure: Comp = TP/(TP+FN) The fitness function combines the CF and Comp measures: Fitness = CF  Comp. An initial population is created consisting of randomly generated rules. The process of generating a new population based on prior populations of rules continues until a population, P, evolves where each rule in P satisfies a prespecified fitness threshold.

10 Reference A. A. Freitas, A Survey of Evolutionary Algorithms for Data Mining and Knowledge, in: Advances in Evolutionary Computing, Springer, 2003.