Presented by: Dr Beatriz de la Iglesia

Slides:



Advertisements
Similar presentations
Introduction to Computer Science 2 Lecture 7: Extended binary trees
Advertisements

CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Chapter 4: Trees Part II - AVL Tree
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
1 Test-Cost Sensitive Naïve Bayes Classification X. Chai, L. Deng, Q. Yang Dept. of Computer Science The Hong Kong University of Science and Technology.
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
Non-Linear Problems General approach. Non-linear Optimization Many objective functions, tend to be non-linear. Design problems for which the objective.
Data Mining Techniques Outline
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Reduced Support Vector Machine
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
Genetic Algorithm.
On comparison of different approaches to the stability radius calculation Olga Karelkina Department of Mathematics University of Turku MCDM 2011.
A Comparison of Nature Inspired Intelligent Optimization Methods in Aerial Spray Deposition Management Lei Wu Master’s Thesis Artificial Intelligence Center.
GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Learning by Simulating Evolution Artificial Intelligence CSMC February 21, 2002.
Mining Binary Constraints in Feature Models: A Classification-based Approach Yi Li.
1 Appendix D: Application of Genetic Algorithm in Classification Duong Tuan Anh 5/2014.
Kanpur Genetic Algorithms Laboratory IIT Kanpur 25, July 2006 (11:00 AM) Multi-Objective Dynamic Optimization using Evolutionary Algorithms by Udaya Bhaskara.
Christoph Eick: Learning Models to Predict and Classify 1 Learning from Examples Example of Learning from Examples  Classification: Is car x a family.
EE749 I ntroduction to Artificial I ntelligence Genetic Algorithms The Simple GA.
2/29/20121 Optimizing LCLS2 taper profile with genetic algorithms: preliminary results X. Huang, J. Wu, T. Raubenhaimer, Y. Jiao, S. Spampinati, A. Mandlekar,
Genetic Algorithms based Parameter Identification of Yeast Fed-Batch Cultivation Angelova M., Tzonkov St., Pencheva T.
Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Multiobjective Optimization  Particularities of multiobjective optimization  Multiobjective.
Evolutionary multi-objective algorithm design issues Karthik Sindhya, PhD Postdoctoral Researcher Industrial Optimization Group Department of Mathematical.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Data Mining CH6 Implementation: Real machine learning schemes(2) Reporter: H.C. Tsai.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
Genetic Programming. What is Genetic Programming? GP for Symbolic Regression Other Representations for GP Example of GP for Knowledge Discovery Outline.
Genetic Algorithms. Solution Search in Problem Space.
Estimation of Distribution Algorithm and Genetic Programming Structure Complexity Lab,Seoul National University KIM KANGIL.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Power Magnetic Devices: A Multi-Objective Design Approach
Machine Learning in Practice Lecture 18
Purposeful Model Parameters Genesis in Simple Genetic Algorithms
Selected Topics in CI I Genetic Programming Dr. Widodo Budiharto 2014.
Evolving Decision Rules (EDR)
An Evolutionary Approach
Alan P. Reynolds*, David W. Corne and Michael J. Chantler
An Empirical Comparison of Supervised Learning Algorithms
USING MICROBIAL GENETIC ALGORITHM TO SOLVE CARD SPLITTING PROBLEM.
Computational Intelligence: Methods and Applications
Iterative Deepening A*
Bulgarian Academy of Sciences
Rule Induction for Classification Using
C.-S. Shieh, EC, KUAS, Taiwan
MultiRefactor: Automated Refactoring To Improve Software Quality
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Example: Applying EC to the TSP Problem
Data Mining (and machine learning)
Decision Tree Saed Sayad 9/21/2018.
Artificial Intelligence Problem solving by searching CSC 361
Heuristic-based Recommendation for Metamodel–OCL Coevolution
Genetic Algorithms: A Tutorial
Example: Applying EC to the TSP Problem
Multi-Objective Optimization
Introduction to Artificial Intelligence Lecture 9: Two-Player Games I
○ Hisashi Shimosaka (Doshisha University)
Statistical Learning Dong Liu Dept. EEIS, USTC.
Learning Chapter 18 and Parts of Chapter 20
Traveling Salesman Problem by Genetic Algorithm
Evolutionary Ensembles with Negative Correlation Learning
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
Beyond Classical Search
Alex Bolsoy, Jonathan Suggs, Casey Wenner
- Hi :D Andrew Lensen, Dr. Bing Xue, and Prof. Mengjie Zhang
Genetic Algorithms: A Tutorial
Multiobjective Optimization
Presentation transcript:

Presented by: Dr Beatriz de la Iglesia Managing Population Diversity Through the Use of Weighted Objectives and Modified Dominance: An Example from Data Mining Alan P. Reynolds and Beatriz de la Iglesia School of Computing Sciences University of East Anglia Norwich, Norfolk, UK Presented by: Dr Beatriz de la Iglesia MCDM 2007

Overview The context: Multi-objective predictive classification The problem: Loss of diversity Managing diversity using three objectives Managing diversity using adaptive objectives Managing diversity by modifying the dominance relation Conclusions and further research Acknowledgement: Work supported by EPSRC Grant GR/T04298 B de la Iglesia MCDM 2007

Multi-Objective (MO) Predictive Classification NSGA II applied to optimize rules in the form of expression trees. Internal nodes are Boolean operators (‘AND’, ‘OR’), leaf nodes contain Attribute Tests (ATs) (e.g. sex=male, color  red, age 29). Expression trees used to predict class in binary classification problems. B de la Iglesia MCDM 2007

Objectives for MO approach Minimizing misclassification costs: simple error rate; balanced error rate; any other measure of overall misclassification cost. Minimizing rule complexity: simple count of ATs or other measures of complexity; to encourage the production of rules that are easily understood by the client; to reduce the chance of overfitting the data. B de la Iglesia MCDM 2007

Applying NSGA II Initialization: population initialized with random balanced trees of depth two. Crossover: subtree crossover which selects two nodes at random and swaps associated subtrees. Mutation applied whenever crossover is not applied 50% chance of AT mutated; 25% chance random AT and its parent node removed 25% chance random AT and parent node inserted B de la Iglesia MCDM 2007

Rule evaluation Rule Simplification: effort is made before each rule is evaluated to simplify the rule. Rule Reduction: If, after rule simplification, the rule is still larger than a preset maximum size (20 ATs in the experiments reported here), ATs and their parent nodes are removed at random until the rule meets the size constraint. B de la Iglesia MCDM 2007

Using two objectives Validation performance for UCI ML repository Adult dataset Population of 100 and crossover rate of 30%. Algorithm run 30 times, error bars give Standard Deviation 14.45% error rate on test data (STD 0.12%) 15.98 % error rate on test data (STD 0.70%) B de la Iglesia MCDM 2007

Loss of population diversity Analysis of performance reveals loss of population diversity (genetic material) early in the search. Information entropy used as a measure of population diversity Entropy is 0 if all solutions are the same in population If all solutions are unique B de la Iglesia MCDM 2007

Population entropy B de la Iglesia MCDM 2007

Population Diversity in MO GAs Early in the search few solutions are non-dominated. Algorithms like NSGA II and SPEA 2 become too elitist. May also be a problem, especially early on in the search, whenever objectives are correlated in other MO scenarios. Solution: artificially increase the fraction of non- dominated solutions early in the search to counteract the loss of population diversity. B de la Iglesia MCDM 2007

Using three objectives If the client is unsure how to calculate misclassification costs, it is tempting to optimise rule complexity, number of False Positives (FP) and False Negatives (FN). In practice, the client usually a rough idea – this knowledge can be used to improve the search. Rule complexity  FP + (1- ) FN (1- ) FP +  FN False positives False negatives Dominated area 3 basic objectives v. 3 objectives with =0.8 at constant rule complexity B de la Iglesia MCDM 2007

Optimizing 3 objectives- 200 generations Results after 200 generations, scaled with respect to =0.5 =0 equivalent to optimizing 3 objectives (Complexity, FP,FN) =0.5 equivalent to optimizing 2 objectives (Complexity, simple error rate) B de la Iglesia MCDM 2007

Optimizing 3 objectives- 2000 generations Results after 2000 generations, scaled with respect to =0.5 including 95% confidence intervals Two objective approach outperforms others, except for =0.4 B de la Iglesia MCDM 2007

Population Diversity Three objectives reduce loss of diversity early in the search Three objectives’ search does not find best large rules under error rate and rule simplicity Solution: Combination approach should be used B de la Iglesia MCDM 2007

Adaptive objectives (2000 gen) Initialise =0. Measuring entropy of population at each iteration, increase  by 0.01 up to 0.5, whenever entropy is greater than 5 and decrease by 0.01 up to 0 whenever entropy less than three. Mean error rates (scaled) on training data using 2000 generations B de la Iglesia MCDM 2007

Adaptive objectives (200 gen) 15.98 % error rate on test data (STD 0.70%) 15.00 % error rate on test data (STD 0.25%) B de la Iglesia MCDM 2007

Modifying the dominance relation Breaking one objective into two or more parts not always easy. An alternative is to use linear combinations of objectives as before, but where one of the weights is negative. For example, minimize 1.1 * complexity – 0.1 * cost 1.1 * cost – 0.1 * complexity Problem: All 1 AT rules become non-dominated. B de la Iglesia MCDM 2007

Modifying dominance relation Alternative is to modify dominance relation by allowing a certain amount of leeway; r1 dominates r2 iff Opposite to -dominance which allows a solution to dominate more of the objective space to control spread of Pareto-front or external archive. B de la Iglesia MCDM 2007

Modifying dominance Results are similar to those obtained when using three objectives in terms of Solution quality Effect on population diversity Results also suggest an adaptive scheme whereby early in the search more leeway is given and this is reduced later. Results of adaptive leeway also provide improvements for both 200 and 2000 generations, although less marked than when using three adaptive objectives. B de la Iglesia MCDM 2007

Conclusions Expression trees optimisation – loss of population diversity early in the search process as small rule with low cost dominates population. This can also be expected in MO problems with few positively correlated objectives. Two methods for maintaining population diversity: Split objective into component parts and use linear combinations of new objectives Modify dominance relation to give leeway in one objective Parameters result in feedback control mechanism depending on diversity in the population B de la Iglesia MCDM 2007

Further research Further experimentation required crossover, population size feedback mechanism other measures of population diversity Effects of diversity loss in other problem domains should be studied. B de la Iglesia MCDM 2007