Optimal n fe Tian-Li Yu & Kai-Chun Fan. n fe n fe = Population Size × Convergence Time n fe is one of the common used metrics to measure the performance.

Slides:



Advertisements
Similar presentations
QoS-based Management of Multiple Shared Resources in Dynamic Real-Time Systems Klaus Ecker, Frank Drews School of EECS, Ohio University, Athens, OH {ecker,
Advertisements

Lindsey Bleimes Charlie Garrod Adam Meyerson
Design of Experiments Lecture I
Roundoff and truncation errors
Size-estimation framework with applications to transitive closure and reachability Presented by Maxim Kalaev Edith Cohen AT&T Bell Labs 1996.
AVL Trees COL 106 Amit Kumar Shweta Agrawal Slide Courtesy : Douglas Wilhelm Harder, MMath, UWaterloo
CSE 373: Data Structures and Algorithms Lecture 5: Math Review/Asymptotic Analysis III 1.
Fundamentals of Python: From First Programs Through Data Structures
Outline. Theorem For the two processor network, Bit C(Leader) = Bit C(MaxF) = 2[log 2 ((M + 2)/3.5)] and Bit C t (Leader) = Bit C t (MaxF) = 2[log 2 ((M.
MATH 685/ CSI 700/ OR 682 Lecture Notes
RANSAC experimentation Slides by Marc van Kreveld 1.
Chapter 8 Estimating Single Population Parameters
Introduction to Analysis of Algorithms
458 Interlude (Optimization and other Numerical Methods) Fish 458, Lecture 8.
Chapter 4: Divide and Conquer The Design and Analysis of Algorithms.
A new crossover technique in Genetic Programming Janet Clegg Intelligent Systems Group Electronics Department.
Quicksort.
DAST, Spring © L. Joskowicz 1 Data Structures – LECTURE 1 Introduction Motivation: algorithms and abstract data types Easy problems, hard problems.
Solver & Optimization Problems n An optimization problem is a problem in which we wish to determine the best values for decision variables that will maximize.
Improved results for a memory allocation problem Rob van Stee University of Karlsruhe Germany Leah Epstein University of Haifa Israel WADS 2007 WAOA 2007.
Genetic Programming.
Roots of Equations Chapter 3. Roots of Equations Also called “zeroes” of the equation –A value x such that f(x) = 0 Extremely important in applications.
Scientific Computing Algorithm Convergence and Root Finding Methods.
Towers of Hanoi. Introduction This problem is discussed in many maths texts, And in computer science an AI as an illustration of recursion and problem.
Frame by Frame Bit Allocation for Motion-Compensated Video Michael Ringenburg May 9, 2003.
Dr. Richard Young Optronic Laboratories, Inc..  Uncertainty budgets are a growing requirement of measurements.  Multiple measurements are generally.
Lecture 8. How to Form Recursive relations 1. Recap Asymptotic analysis helps to highlight the order of growth of functions to compare algorithms Common.
A Sampling Distribution
Estimation of Statistical Parameters
1 Statistical Inference Greg C Elvers. 2 Why Use Statistical Inference Whenever we collect data, we want our results to be true for the entire population.
Statistics 101 Chapter 10. Section 10-1 We want to infer from the sample data some conclusion about a wider population that the sample represents. Inferential.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Unit III : Introduction To Data Structures and Analysis Of Algorithm 10/8/ Objective : 1.To understand primitive storage structures and types 2.To.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
© Negnevitsky, Pearson Education, Lecture 10 Evolutionary Computation: Evolution strategies and genetic programming Evolution strategies Evolution.
A Polynomial Time Approximation Scheme For Timing Constrained Minimum Cost Layer Assignment Shiyan Hu*, Zhuo Li**, Charles J. Alpert** *Dept of Electrical.
Genetic Algorithms Michael J. Watts
Télécom 2A – Algo Complexity (1) Time Complexity and the divide and conquer strategy Or : how to measure algorithm run-time And : design efficient algorithms.
Accuracy Based Generation of Thermodynamic Properties for Light Water in RELAP5-3D 2010 IRUG Meeting Cliff Davis.
A Study of Balanced Search Trees: Brainstorming a New Balanced Search Tree Anthony Kim, 2005 Computer Systems Research.
Section 10.1 Confidence Intervals
O PTIMAL SERVICE TASK PARTITION AND DISTRIBUTION IN GRID SYSTEM WITH STAR TOPOLOGY G REGORY L EVITIN, Y UAN -S HUN D AI Adviser: Frank, Yeong-Sung Lin.
Kanpur Genetic Algorithms Laboratory IIT Kanpur 25, July 2006 (11:00 AM) Multi-Objective Dynamic Optimization using Evolutionary Algorithms by Udaya Bhaskara.
Java Methods Big-O Analysis of Algorithms Object-Oriented Programming
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
The Bisection Method. Introduction Bisection Method: Bisection Method = a numerical method in Mathematics to find a root of a given function.
Energy-Aware Scheduling for Aperiodic Tasks on Multi-core Processors Dawei Li and Jie Wu Department of Computer and Information Sciences Temple University,
Rounding scheme if r * j  1 then r j := 1  When the number of processors assigned in the continuous solution is between 0 and 1 for each task, the speed.
GENETIC ALGORITHM Basic Algorithm begin set time t = 0;
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
A Different Solution  alternatively we can use the following algorithm: 1. if n == 0 done, otherwise I. print the string once II. print the string (n.
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
MultiModality Registration Using Hilbert-Schmidt Estimators By: Srinivas Peddi Computer Integrated Surgery II April 6 th, 2001.
1 Probability and Statistics Confidence Intervals.
Tutorial 2, Part 2: Calibration of a damped oscillator.
Selection Methods Choosing the individuals in the population that will create offspring for the next generation. Richard P. Simpson.
Program Performance 황승원 Fall 2010 CSE, POSTECH. Publishing Hwang’s Algorithm Hwang’s took only 0.1 sec for DATASET1 in her PC while Dijkstra’s took 0.2.
Clustering Data Streams A presentation by George Toderici.
The normal approximation for probability histograms.
Ch03-Algorithms 1. Algorithms What is an algorithm? An algorithm is a finite set of precise instructions for performing a computation or for solving a.
Chapter 15 Running Time Analysis. Topics Orders of Magnitude and Big-Oh Notation Running Time Analysis of Algorithms –Counting Statements –Evaluating.
CSE 3358 NOTE SET 2 Data Structures and Algorithms 1.
Hirophysics.com The Genetic Algorithm vs. Simulated Annealing Charles Barnes PHY 327.
CSE 330: Numerical Methods. What is true error? True error is the difference between the true value (also called the exact value) and the approximate.
Algorithm Analysis (not included in any exams!)
Objective of This Course
○ Hisashi Shimosaka (Doshisha University)
PERFORMANCE MEASURES. COMPUTATIONAL MODELS Equal Duration Model:  It is assumed that a given task can be divided into n equal subtasks, each of which.
Presentation transcript:

Optimal n fe Tian-Li Yu & Kai-Chun Fan

n fe n fe = Population Size × Convergence Time n fe is one of the common used metrics to measure the performance of GAs. The optimal n fe is an n fe that is based on a reliable population size (large enough to get good solutions.) This slide propose a new method, called Optimal n fe, to find the optimal n fe.

Agenda User guide Bisection Optimal n fe Conclusion

Installation Components of DSMGA Source codes only – Modified files (they do not affect the original programs) Makefile global.cpp global.h – Additional file optnfe.cpp After compiling – optnfe

Usage./optnfe ell lower upper – optnfe executable file – ell problem size – lower population size lower bound -1 for no assignment – upper population size upper bound -1 for no assignment

Algorithm 2 x 2 phases – Bisection ( 2 phases ) Lower bound of reliable population size to solve problem – Optimal n fe ( 2 phases) Optimal (minimal) n fe to solve problem

Agenda User guide Bisection Optimal n fe Conclusion

Bisection Method in Mathematics A root-finding algorithm which repeatedly bisects an interval then selects a subinterval in which a root must lie for further processing Intermediate value theorem f:[a, b] → R, if f(a) < u < f(b) or f(b) < u < f(a), then there exists c in (a, b) where f(c) = u.

Bisection Method in GAs Martin Pelikan, 2005 – Given that we know the optimum (population size), the number of generations can be determined by the GA itself. – The problem is how to determine an adequate population size for each problem size. This slide uses “Bisection” to call the bisection method proposed by Martin Pelikan.

Bisection (Martin Pelikan) 1. Start with some very small N (say, N = 10 ) 2. Double N until GA convergence is sufficiently reliable 3. (min, max) = (N/2, N) 4. repeat until (max - min) / min < 0.1 (use your own threshold here) { N = (min + max) / 2 if N leads to sufficiently reliable convergence then max = N else min = N } 5. Compute all your statistics for this problem size using pop.size = max

Bisection in GAs Martin Pelikan: What I like on using the bisection method is that all parameter tweaking will be done automatically, because the remaining GA parameters do not influence time complexity as much (mostly by some constant factor and that's not so important when we talk about asymptotic upper bounds). To eliminate noise, the bisection can be run several times and the results of all bisection runs can be averaged (right now I use 10 rounds of bisection, each with 10 runs).

Bisection in GAs Martin Pelikan: Of course, the above thinking is correct only if larger populations lead to better solutions (given enough generations we should all be able to believe this fact) and if smaller populations do not significantly increase the number of generations (in all cases I studied, any sufficiently large population size lead to about the same number of generations, as can be expected for many variants of evolutionary algorithms based primarily on recombination).

Population Size vs. Convergence Time n fe = Population Size × Convergence Time – Order Population Size > Convergence Time – Constant? population size lower bound (optimum) found by Bisection with 10×100 runs in the experiments for DMC n fe population size

Agenda User guide Bisection Optimal n fe Conclusion

Empirical results – According to the abilities of GAs? DMC(left), perfect model for trap(right) Maybe the scale of lower bound is also a key point. Assumption Population size lower bound n fe population size

Prerequisite Population size lower bound from Bisection – Efficiency – An accurate search for n fe may take thousands of runs for each population size. – Empirical results show that meaningful search ranges are not so large. Touching the impractical population size (too small or to large) may waste lots of time.

Optimal n fe 2 phases – Phase 1 Calculate the n fe of population size lower bound (from Bisection). Search for the range (upper bound) of population size according to n fe. – Phase 2 Divide the range into four pieces, and then calculate the slope for each one. Reduce the range recursively by observing the variations in slopes.

Optimal n fe Phase 1 Case 1 – n fe(1) < n fe(2) n fe population size n fe(1) n fe(2) some ratio (10% population size lower bound in default)

Optimal n fe Phase 1 Case 2 – n fe(1) > n fe(2) AND n fe(3) > n fe(2) n fe population size n fe(1) n fe(2) n fe(1) n fe(2) Which one? n fe(3)

Optimal n fe Phase 1 Case 3 (extension of case 2) – n fe(1) > n fe(2) > … > n fe(i-1) – n fe(i-1) > n fe(i) AND n fe(i+1) > n fe(i) n fe population size n fe(1) n fe(5) n fe(6) n fe(2) n fe(3) n fe(4) like case 2

Optimal n fe Phase 2 7 (or 9) cases – Case 1 (and 8) n fe(2) - n fe(1) ≥ 0

Optimal n fe Phase 2 7 cases – Case 2 & n fe(2) - n fe(1) < 0 n fe(3) - n fe(2) > 0 n fe(3) - n fe(2) = 0

Optimal n fe Phase 2 7 cases – Case 4 & n fe(3) - n fe(2) < 0 n fe(4) - n fe(3) > 0 n fe(4) - n fe(3) = 0

Optimal n fe Phase 2 7 cases – Case 6 & 7 (and 9) n fe(4) - n fe(3) < 0 n fe(5) - n fe(4) > 0 n fe(5) - n fe(4) ≤

Optimal n fe Algorithm 1.Start with the optimum population size pop (0) from Bisection and then get n fe(0) 2.Calculate n fe(i+1) where pop (i+1) = pop (i) *1.1, repeat until n fe(i+1) > n fe(i) 3.Keep the information of i-1 as lower and i+1 as upper 4.Repeat until pop (upper) – pop (lower) < pop (0) * 0.01 (to assign the precision here) { for ( k = 1 to 4 ) { calculate n fe of pop (k), where pop (k) = ( ((4 – k) * pop (lower) + k * pop (upper) ) / 4 ) if n fe(k) ≥ n fe(k-1) { lower ← k - 2 or 0 (if k = 1) upper ← k } 5.Optimal n fe is the minimum n fe with its population size between pop (lower) and pop (upper) that has already been calculated in above runs.

Speed Up Unnecessary calculation skipping – If and only if n fe(i) - n fe(i-1) < 0, then n fe(i+1) - n fe(i) should be considered. Table look-up

In fact Empirical results show that, the population size difference between Bisection and Optimal n fe are almost lower than 20%. – For example, if the Bisection optimum is 1000, the population size for optimal n fe is smaller than – If the precision is 1%, the worst case of sweep method is 20 runs, the average case is less than 10 runs. – However, if higher precision is needed, Optimal n fe beats the sweep method obviously. 1% (100 for 10000): 10 runs vs. 11 runs 0.1% (10 for 10000): 100 runs vs. 17 runs 0.01% (1 for 10000): 1000 runs vs. 25runs

Agenda User guide Bisection Optimal n fe Conclusion

Problems and Future Work Some empirical results show that n fe over population size may be a concave-up or an increasing curve. – A proof or strict experiments are needed. – Due to the abilities of GAs? Maybe the more powerful the GA is, the closer results are from Optimal n fe to Bisection. Maybe the scale of lower bound is a key point. If there exist a problem such that Optimal n fe and Bisection result in contraries?

Conclusion n fe is one of the commonly used metrics to measure the performance of GAs. Bisection is a method to find the minimum reliable population size efficiently. Due to some empirical results, the optima from Bisection are not necessarily the population sizes for optimal n fe. This slide propose a new method, Optimal n fe, which can automatically calculate the optimal (minimal) n fe for GAs processing. The difference between Bisection and Optimal n fe still has to be discussed in the future.