Genetic Programming. GP quick overview Developed: USA in the 1990’s Early names: J. Koza Typically applied to: machine learning tasks (prediction, classification…)

Slides:



Advertisements
Similar presentations
Biologically Inspired AI (mostly GAs). Some Examples of Biologically Inspired Computation Neural networks Evolutionary computation (e.g., genetic algorithms)
Advertisements

Genetic Algorithms Representation of Candidate Solutions GAs on primarily two types of representations: –Binary-Coded –Real-Coded Binary-Coded GAs must.
Genetic Programming 김용덕 Page 2 Contents What is Genetic Programming? Difference between GP and GA Flowchart for GP Structures in GP.
Evolutionary Computational Intelligence
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Doug Downey, adapted from Bryan Pardo, Machine Learning EECS 349 Machine Learning Genetic Programming.
Intro to AI Genetic Algorithm Ruth Bergman Fall 2002.
Evolutionary Computational Intelligence
Chapter 6: Transform and Conquer Genetic Algorithms The Design and Analysis of Algorithms.
Genetic Programming. Agenda What is Genetic Programming? Background/History. Why Genetic Programming? How Genetic Principles are Applied. Examples of.
Genetic Programming Dinesh Dharme Prateek Srivastav Pankhil Chheda
Genetic Algorithms Overview Genetic Algorithms: a gentle introduction –What are GAs –How do they work/ Why? –Critical issues Use in Data Mining –GAs.
Genetic Programming.
Genetic Programming Chapter 6. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Genetic Programming GP quick overview Developed: USA.
Medical Diagnosis via Genetic Programming Project #2 Artificial Intelligence: Biointelligence Computational Neuroscience Connectionist Modeling of Cognitive.
Genetic Algorithm.
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Evolution Strategies Evolutionary Programming Genetic Programming Michael J. Watts
Genetic algorithms Prof Kang Li
Automatically Defined Functions Used to evolve modular programs Architecture implemented by Koza Architecture implemented by Bruce Function calls A critical.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
Design of an Evolutionary Algorithm M&F, ch. 7 why I like this textbook and what I don’t like about it!
Introduction to Evolutionary Algorithms Session 4 Jim Smith University of the West of England, UK May/June 2012.
Biological data mining by Genetic Programming AI Project #2 Biointelligence lab Cho, Dong-Yeon
Genetic Algorithms Siddhartha K. Shakya School of Computing. The Robert Gordon University Aberdeen, UK
Brief introduction to genetic algorithms and genetic programming A.E. Eiben Free University Amsterdam.
G ENETIC P ROGRAMMING Ranga Rodrigo March 17,
Artificial Intelligence Chapter 4. Machine Evolution.
Initial Population Generation Methods for population generation: Grow Full Ramped Half-and-Half Variety – Genetic Diversity.
1 Genetic Algorithms K.Ganesh Introduction GAs and Simulated Annealing The Biology of Genetics The Logic of Genetic Programmes Demo Summary.
Chapter 9 Genetic Algorithms.  Based upon biological evolution  Generate successor hypothesis based upon repeated mutations  Acts as a randomized parallel.
Evolutionary Programming
Evolutionary Computing
Evolutionary Computing Dialects Presented by A.E. Eiben Free University Amsterdam with thanks to the EvoNet Training Committee and its “Flying Circus”
Genetic Algorithms What is a GA Terms and definitions Basic algorithm.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
Project 2: Classification Using Genetic Programming Kim, MinHyeok Biointelligence laboratory Artificial.
Chapter 12 FUSION OF FUZZY SYSTEM AND GENETIC ALGORITHMS Chi-Yuan Yeh.
2101INT – Principles of Intelligent Systems Lecture 11.
Genetic Programming A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 6.
Genetic Algorithms. The Basic Genetic Algorithm 1.[Start] Generate random population of n chromosomes (suitable solutions for the problem) 2.[Fitness]
GENETIC PROGRAMMING. THE CHALLENGE "How can computers learn to solve problems without being explicitly programmed? In other words, how can computers be.
Automated discovery in math Machine learning techniques (GP, ILP, etc.) have been successfully applied in science Machine learning techniques (GP, ILP,
RMIT UNIVERSITY ASPGPAnalysis of Genetic Programming Runs1 Vic Ciesielski and Xiang Li {vc, School of Computer Science and Information.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Genetic Programming COSC Ch. F. Eick, Introduction to Genetic Programming GP quick overview Developed: USA in the 1990’s Early names: J. Koza Typically.
John R. Koza [Edited by J. Wiebe] 1. GENETIC PROGRAMMING 2.
Symbolic Regression via Genetic Programming AI Project #2 Biointelligence lab Cho, Dong-Yeon
Genetic Programming Using Simulated Natural Selection to Automatically Write Programs.
1 Comparative Study of two Genetic Algorithms Based Task Allocation Models in Distributed Computing System Oğuzhan TAŞ 2005.
Evolutionary Programming A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 5.
Genetic Programming. What is Genetic Programming? GP for Symbolic Regression Other Representations for GP Example of GP for Knowledge Discovery Outline.
A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolutionary Programming.
Genetic Programming.
Genetic Algorithms Author: A.E. Eiben and J.E. Smith
Evolutionary Programming
Evolution Strategies Evolutionary Programming
Medical Diagnosis via Genetic Programming
Machine Learning Evolutionary Algorithm (2)
Optimization and Learning via Genetic Programming
Artificial Intelligence Chapter 4. Machine Evolution
Genetic Algorithms Chapter 3.
Artificial Intelligence Chapter 4. Machine Evolution
Evolutionary Programming
Genetic Programming Chapter 6.
Genetic Programming.
Genetic Programming Chapter 6.
Genetic Programming Chapter 6.
Beyond Classical Search
Presentation transcript:

Genetic Programming

GP quick overview Developed: USA in the 1990’s Early names: J. Koza Typically applied to: machine learning tasks (prediction, classification…) Attributed features: competes with neural nets and alike needs huge populations (thousands) slow Special: non-linear chromosomes: trees, graphs mutation possible but not necessary (disputed!)

Introductory example: credit scoring Bank wants to distinguish good from bad loan applicants Model needed that matches historical data IDNo of children SalaryMarital status OK? ID Married 0 ID Single 1 ID Divorced 1 …

Introductory example: credit scoring A possible model: IF (NOC = 2) AND (S > 80000) THEN good ELSE bad In general: IF formula THEN good ELSE bad Only unknown is the right formula, hence Our search space (phenotypes) is the set of formulas Natural fitness of a formula: percentage of well classified cases of the model it stands for Natural representation of formulas (genotypes) is: parse trees

Introductory example: credit scoring IF (NOC = 2) AND (S > 80000) THEN good ELSE bad can be represented by the following tree AND S2NOC >=

Tree based representation Trees are a universal form, e.g. consider Arithmetic formula Logical formula Program (x  true)  (( x  y )  (z  (x  y))) i =1; while (i < 20) { i = i +1 }

Tree based representation

i =1; while (i < 20) { i = i +1 }

Tree based representation In GA, ES, EP chromosomes are linear structures (bit strings, integer string, real- valued vectors, permutations) Tree shaped chromosomes are non-linear structures In GA, ES, EP the size of the chromosomes is fixed Trees in GP may vary in depth and width

Tree based representation Symbolic expressions can be defined by Terminal set T Function set F (with the arities of function symbols) Adopting the following general recursive definition: 1. Every t  T is a correct expression 2. f(e 1, …, e n ) is a correct expression if f  F, arity(f)=n and e 1, …, e n are correct expressions 3. There are no other forms of correct expressions

GP flowchartGA flowchart

Mutation Most common mutation: replace randomly chosen subtree by randomly generated tree

Mutation cont’d Mutation has two parameters: Probability p m to choose mutation vs. recombination Probability to chose an internal point as the root of the subtree to be replaced Remarkably p m is advised to be 0 (Koza’92) or very small, like 0.05 (Banzhaf et al. ’98) The size of the child can exceed the size of the parent

Recombination Most common recombination: exchange two randomly chosen subtrees among the parents Recombination has two parameters: Probability p c to choose recombination vs. mutation Probability to chose an internal point within each parent as crossover point The size of offspring can exceed that of the parents

Child 2 Parent 1Parent 2 Child 1

Selection Parent selection typically fitness proportionate Over-selection in very large populations rank population by fitness and divide it into two groups: group 1: best x% of population, group 2 other (100-x)% 80% of selection operations chooses from group 1, 20% from group 2 for pop. size = 1000, 2000, 4000, 8000 x = 32%, 16%, 8%, 4% motivation: to increase efficiency, %’s come from rule of thumb Survivor selection: Typical: generational scheme (thus none) Recently steady-state is becoming popular for its elitism

Initialisation Maximum initial depth of trees D max is set Full method (each branch has depth = D max ): nodes at depth d < D max randomly chosen from function set F nodes at depth d = D max randomly chosen from terminal set T Grow method (each branch has depth  D max ): nodes at depth d < D max randomly chosen from F  T nodes at depth d = D max randomly chosen from T Common GP initialisation: ramped half-and-half, where grow & full method each deliver half of initial population

Bloat Bloat = “survival of the fattest”, i.e., the tree sizes in the population are increasing over time Ongoing research and debate about the reasons Needs countermeasures, e.g. Prohibiting variation operators that would deliver “too big” children Parsimony pressure: penalty for being oversized

Problems involving “physical” environments Trees for data fitting vs. trees (programs) that are “really” executable Execution can change the environment  the calculation of fitness Example: robot controller Fitness calculations mostly by simulation, ranging from expensive to extremely expensive (in time) But evolved controllers are often to very good

Example application: symbolic regression Given some points in R 2, (x 1, y 1 ), …, (x n, y n ) Find function f(x) s.t.  i = 1, …, n : f(x i ) = y i Possible GP solution: Representation by F = {+, -, /, sin, cos}, T = R  {x} Fitness is the error All operators standard pop.size = 1000, ramped half-half initialisation Termination: n “hits” or fitness evaluations reached (where “hit” is if | f(x i ) – y i | < )

Discussion Is GP: The art of evolving computer programs ? Means to automated programming of computers? GA with another representation?

Introduction to GP based identification System Model Black-box models Model-selection Black-box modeling A modellező választ modell struktúrát Model structureModel parameters Known parameters Based on prior knowledge Identification From measured data with some optimization

Linear in parameters model The original GP generates models that are nonlinear in their parameters Linear in parameters model The parameters can be estimated by LS

Model representation + + * x2x2 x1x1 x1x1 + * x2x2 x1x1 x3x3 F1F1 F2F2 F3F3

OLS 1.With orthogonal LS we can estimate the contribution 2.The useless branches will be deleted + + * x2x2 x1x1 x1x1 + * x2x2 x1x1 x3x3 F1F1 F2F2 F3F3 + * x2x2 x1x1 x1x1 F1F1 F2F2

Example: system-identification (u) (y) + 4% noisesimulation Result of GP

Improvement of the OLS Method123 Not exact solution from exact solution from Avarage runtime to find a solution runs with max evaluations 1.: No penaly on the size of the tree 2.: Penaly on the size of the tree 3.: penaly on the size of the tree + OLS

Conclusions Model structure identification is a not trivial task GP can be used for this purpose Linear in parameters models are ideal for GP based structure exploration It is useful to apply OLS to regularize the tree