Fast Evolutionary Optimisation Temi avanzati di Intelligenza Artificiale - Lecture 6 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.

Slides:



Advertisements
Similar presentations
Temi Avanzati di Intelligenza Artificiale - Intro1 Temi Avanzati di Intelligenza Artificiale Prof. Vincenzo Cutello Department of Mathematics and Computer.
Advertisements

Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
On the Genetic Evolution of a Perfect Tic-Tac-Toe Strategy
Continuous Probability Distributions.  Experiments can lead to continuous responses i.e. values that do not have to be whole numbers. For example: height.
1 Simulation Modeling and Analysis Session 13 Simulation Optimization.
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Tutorial 2 (Lecture6) (Intro to) Evolutionary Computation School of Computer Science University of Birmingham.
A simple EA and Common Search Operators Temi avanzati di Intelligenza Artificiale - Lecture 2 Prof. Vincenzo Cutello Department of Mathematics and Computer.
Evaluating Hypotheses
An Introduction to Black-Box Complexity
Tutorial 2 Temi avanzati di Intelligenza Artificiale - Lecture 6 Prof. Vincenzo Cutello Department of Mathematics and Computer Science University of Catania.
Evolutionary Computational Intelligence Lecture 9: Noisy Fitness Ferrante Neri University of Jyväskylä.
Tutorial 1 Temi avanzati di Intelligenza Artificiale - Lecture 3 Prof. Vincenzo Cutello Department of Mathematics and Computer Science University of Catania.
Evolutionary Computational Intelligence
CHAPTER 6 Statistical Analysis of Experimental Data
Continuous Random Variables and Probability Distributions
Properties of Random Numbers
Genetic Algorithm What is a genetic algorithm? “Genetic Algorithms are defined as global optimization procedures that use an analogy of genetic evolution.
Tutorial 4 (Lecture 12) Remainder from lecture 10: The implicit fitness sharing method Exercises – any solutions? Questions?
Introduction to Simulated Annealing 22c:145 Simulated Annealing  Motivated by the physical annealing process  Material is heated and slowly cooled.
Slides are based on Negnevitsky, Pearson Education, Lecture 10 Evolutionary Computation: Evolution strategies and genetic programming n Evolution.
Genetic Algorithm.
Evolutionary Intelligence
© Negnevitsky, Pearson Education, CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University.
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
Estimation of Statistical Parameters
PROBABILITY & STATISTICAL INFERENCE LECTURE 3 MSc in Computing (Data Analytics)
Using Genetic Programming to Learn Probability Distributions as Mutation Operators with Evolutionary Programming Libin Hong, John Woodward, Ender Ozcan,
Statistical inference. Distribution of the sample mean Take a random sample of n independent observations from a population. Calculate the mean of these.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Constrained Evolutionary Optimization Yong Wang Associate Professor, PhD School of Information Science and Engineering, Central South University
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
© Negnevitsky, Pearson Education, Lecture 10 Evolutionary Computation: Evolution strategies and genetic programming Evolution strategies Evolution.
Schemata Theory Chapter 11. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Theory Why Bother with Theory? Might provide performance.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Neural and Evolutionary Computing - Lecture 6
Simulated Annealing.
Applying Genetic Algorithm to the Knapsack Problem Qi Su ECE 539 Spring 2001 Course Project.
Comp. Genomics Recitation 3 The statistics of database searching.
Statistical Methods II&III: Confidence Intervals ChE 477 (UO Lab) Lecture 5 Larry Baxter, William Hecker, & Ron Terry Brigham Young University.
G5BAIM Artificial Intelligence Methods Dr. Rong Qu Evolutionary Algorithms.
1 A New Method for Composite System Annualized Reliability Indices Based on Genetic Algorithms Nader Samaan, Student,IEEE Dr. C. Singh, Fellow, IEEE Department.
© Negnevitsky, Pearson Education, Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Introduction,
Statistical Methods II: Confidence Intervals ChE 477 (UO Lab) Lecture 4 Larry Baxter, William Hecker, & Ron Terry Brigham Young University.
Method of Hooke and Jeeves
Evolution strategies Luis Martí DEE/PUC-Rio. ES quick overview Developed: Germany in the 1970’s Early names: I. Rechenberg, H.-P. Schwefel Typically applied.
Clustering and Testing in High- Dimensional Data M. Radavičius, G. Jakimauskas, J. Sušinskas (Institute of Mathematics and Informatics, Vilnius, Lithuania)
 Negnevitsky, Pearson Education, Lecture 9 Evolutionary Computation: Genetic algorithms n Introduction, or can evolution be intelligent? n Simulation.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
Probabilistic Algorithms Evolutionary Algorithms Simulated Annealing.
Evolution strategies Chapter 4. A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Evolution Strategies ES quick overview Developed: Germany.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
CHAPTER 2.3 PROBABILITY DISTRIBUTIONS. 2.3 GAUSSIAN OR NORMAL ERROR DISTRIBUTION  The Gaussian distribution is an approximation to the binomial distribution.
Selection Methods Choosing the individuals in the population that will create offspring for the next generation. Richard P. Simpson.
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
Selection and Recombination Temi avanzati di Intelligenza Artificiale - Lecture 4 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Evolutionary Programming A.E. Eiben and J.E. Smith, Introduction to Evolutionary Computing Chapter 5.
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
EVOLUTIONARY SYSTEMS AND GENETIC ALGORITHMS NAME: AKSHITKUMAR PATEL STUDENT ID: GRAD POSITION PAPER.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Virtual University of Pakistan
Virtual University of Pakistan
Geology Geomath Chapter 7 - Statistics tom.h.wilson
Evolutionary Computation,
One-Way Analysis of Variance
Presentation transcript:

Fast Evolutionary Optimisation Temi avanzati di Intelligenza Artificiale - Lecture 6 Prof. Vincenzo Cutello Department of Mathematics and Computer Science University of Catania

Fast Evolutionary Optimisation - Lecture 52 Exercise Sheet  Function Optimisation by Evolutionary Programming  Fast Evolutionary Programming  Computational Studies  Summary

Fast Evolutionary Optimisation - Lecture 53 Global Optimisation

Fast Evolutionary Optimisation - Lecture 54 Global Optimisation by Mutation- Based EAs 1. Generate the initial population of  individuals, and set k = 1. Each individual is a real-valued vector, (x i ). 2. Evaluate the fitness of each individual. 3. Each individual creates a single offspring: for j = 1,…, n, where xi(j) denotes the j-th component of the vectors x i. N(0,1) denotes a normally distributed one-dimensional random number with mean zero and standard deviation one. N j (0,1) indicates that the random number is generated anew for each value of j. 4. Calculate the fitness of each offspring. 5. For each individual, q opponents are chosen randomly from all the parents and offspring with an equal probability. For each comparison, if the individual's fitness is no greater than the opponent's, it receives a “win”.

Fast Evolutionary Optimisation - Lecture Select the  best individuals (from 2  ) that have the most wins to be the next generation. 7. Stop if the stopping criterion is satisfied; otherwise, k = k + 1 and go to Step 3.

Fast Evolutionary Optimisation - Lecture 56 Why N(0,1)?  The standard deviation of the Normal distribution determines the search step size of the mutation. It is a crucial parameter.  Unfortunately, the optimal search step size is problem-dependent.  Even for a single problem, different search stages require different search step sizes.  Self-adaptation can be used to get around this problem partially.

Fast Evolutionary Optimisation - Lecture 57 Function Optimisation by Classical EP (CEP) EP = Evolutionary Programming 1. Generate the initial population of  individuals, and set k = 1. Each individual is taken as a pair of real-valued vectors, (x i,  i ),  i  {1,…,  }. 2. Evaluate the fitness score for each individual (x i,  i ),  i  {1,…,  } of the population based on the objective function, f(x i ). 3. Each parent (x i,  i ), i = 1,…, , creates a single offspring (x’ i,  ’ i ), by: for j = 1,…,n where x j (j), x j ’ (j),  j (j) and  j ’ (j) denote the j-th component of the vectors x j, x j ’,  j and  j ’, respectively. N(0,1) denotes a normally distributed one-dimensional random number with mean zero and standard deviation one. N j (0,1) indicates that the random number is generated anew for each value of j. The factors  and  ’ have commonly set to

Fast Evolutionary Optimisation - Lecture Calculate the fitness of each offspring (x i ’,  i ’ ),  i  {1,…,  }. 5. Conduct pairwise comparison over the union of parents (x i,  i ), and offspring (x i ’,  I ’ ),  i  {1,…,  }. For each individual, q opponents are chosen randomly from all the parents and offspring with an equal probability. For each comparison, if the individual's fitness is no greater than the opponent's, it receives a “win.” 6. Select the  individuals out of (x i,  i ), and (x i ’,  i ’ ),  i  {1,…,  } that have the most wins to be parents of the next generation. 7. Stop if the stopping criterion is satisfied; otherwise, k = k + 1 and go to Step 3.

Fast Evolutionary Optimisation - Lecture 59  What Do Mutation and Self-Adaptation Do

Fast Evolutionary Optimisation - Lecture 510 Fast EP  The idea comes from fast simulated annealing.  Use a Cauchy, instead of Gaussian, random number in Eq.(1) to generate a new offspring. That is, where  j is an Cauchy random number variable with the scale parameter t = 1, and is generated anew for each value of j.  Everything else, including Eq.(2), are kept unchanged in order to evaluate the impact of Cauchy random numbers.

Fast Evolutionary Optimisation - Lecture 511 Cauchy Distribution  Its density function is  where t > 0 is a scale parameter. The corresponding distribution function is

Fast Evolutionary Optimisation - Lecture 512 Gaussian and Cauchy Density Functions

Fast Evolutionary Optimisation - Lecture 513 Test Functions  23 functions were used in our computational studies. They have different characteristics.  Some have a relatively high dimension.  Some have many local optima.

Fast Evolutionary Optimisation - Lecture 514

Fast Evolutionary Optimisation - Lecture 515 Experimental Setup  Population size 100.  Competition size 10 for selection.  All experiments were run 50 times, i.e., 50 trials.  Initial populations were the same for CEP and FEP.

Fast Evolutionary Optimisation - Lecture 516 Experiments on Unimodal Functions  The value of t with 49 degrees of freedom is significant at  = 0,05 by a two-tailed test.

Fast Evolutionary Optimisation - Lecture 517 Discussions on Unimodal Functions  FEP performed better than CEP on f 3 -f 7.  CEP was better for f 1 and f 2.  FEP converged faster, even for f 1 and f 2 (for a long period).

Fast Evolutionary Optimisation - Lecture 518 Experiments on Multimodal Functions f 8 -f 13  The value of t with 49 degrees of freedom is significant at  = 0,05 by a two-tailed test.

Fast Evolutionary Optimisation - Lecture 519 Discussions on Multimodal Functions f 8 -f 13  FEP converged faster to a better solution.  FEP seemed to deal with many local minima well.

Fast Evolutionary Optimisation - Lecture 520 Experiments on Multimodal Functions f 14 -f 23  The value of t with 49 degrees of freedom is significant at  = 0,05 by a two-tailed test.

Fast Evolutionary Optimisation - Lecture 521 Discussions on Multimodal Functions f 14 -f 23  The results are mixed!  FEP and CEP performed equally well on f 16 and f 17. They are comparable on f 15 and f 18 – f 20.  CEP performed better on f 21 – f 23 (Shekel functions).  Is it because the dimension was low so that CEP appeared to be better?

Fast Evolutionary Optimisation - Lecture 522 Experiments on Low-Dimensional f 8 -f 13  The value of t with 49 degrees of freedom is significant at  = 0,05 by a two-tailed test.

Fast Evolutionary Optimisation - Lecture 523 Discussions on Low-Dimensional f 8 -f 13  FEP still converged faster to better solutions.  Dimensionality does not play a major role in causing the difference between FEP and CEP.  There must be something inherent in those functions which caused such difference.

Fast Evolutionary Optimisation - Lecture 524 The Impact of Parameter t on FEP | Part I  For simplicity, t = 1 in all the above experiments for FEP. However, this may not be the optimal choice for a given problem.  Table 1: The mean best solutions found by FEP using different scale parameter t in the Cauchy mutation for functions f 1 (1500), f 2 (2000), f 10 (1500), f 11 (2000), f 21 (100), f 22 (100) and f 23 (100). The values in “()” indicate the number of generations used in FEP. All results have been averaged over 50 runs.

Fast Evolutionary Optimisation - Lecture 525 The Impact of Parameter t on FEP | Part II  Table 2: The mean best solutions found by FEP using different scale parameter t in the Cauchy mutation for functions f 1 (1500), f 2 (2000), f 10 (1500), f 11 (2000), f 21 (100), f 22 (100) and f 23 (100). The values in “()” indicate the number of generations used in FEP. All results have been averaged over 50 runs.

Fast Evolutionary Optimisation - Lecture 526 Why Cauchy Mutation Performed Better  Given G(0,1) and C(1), the expected length of Gaussian and Cauchy jumps are:  It is obvious that Gaussian mutation is much localised than Cauchy mutation.

Fast Evolutionary Optimisation - Lecture 527 Why and When Large Jumps Are Beneficial  (Only one dimensional case is considered here for convenience's sake.)  Take the Gaussian mutation with G(0,  2 ) distribution as an example, i.e.,  the probability of generating a point in the neighbourhood of the global optimum x * is given by :  where  > 0 is the neighbourhood size and  is often regarded as the step size of the Gaussian mutation. Figure 4 illustrates the situation.

Fast Evolutionary Optimisation - Lecture 528  Figure 4: Evolutionary search as neighbourhood search, where x * is the global  optimum and  > 0 is the neighbourhood size.

Fast Evolutionary Optimisation - Lecture 529 An Analytical Result  It can be shown that when x* -  +  > . That is, the larger  is, the larger if x* -  +  > .  On the other hand, if x* -  +  > , then which indicates that decreases, exponentially, as  increases.

Fast Evolutionary Optimisation - Lecture 530 Empirical Evidence I  Table 3: Comparison of CEP's and FEP's final results on f 21 when the initial population is generated uniformly at random in the range of 0  x i  10 and 2.5  x i  5.5. The results were averaged over 50 runs. The number of generations for each run was 100.

Fast Evolutionary Optimisation - Lecture 531 Empirical Evidence II  Table 4: Comparison of CEP's and FEP's final results on f 21 when the initial population is generated uniformly at random in the range of 0  x i  10 and 0  x i  100 and a i 's were multiplied by 10. The results were averaged over 50 runs. The number of generations for each run was 100.

Fast Evolutionary Optimisation - Lecture 532 Summary  Cauchy mutation performs well when the global optimum is far away from the current search location. Its behaviour can be explained theoretically and empirically.  An optimal search step size can be derived if we know where the global optimum is. Unfortunately, such information is unavailable for real-world problems.  The performance of FEP can be improve by a set of more suitable parameters, instead of copying CEP's parameter setting.  Reference X. Yao, Y. Liu and G. Lin, “Evolutionary programming made faster,” IEEE Transactions on Evolutionary Computation, 3(2):82-102, July 1999.