Presentation is loading. Please wait.

Presentation is loading. Please wait.

Numerical Optimization

Similar presentations


Presentation on theme: "Numerical Optimization"— Presentation transcript:

1 Numerical Optimization
General Framework: objective function f(x1,...,xn) to be minimized or maximized constraints: gi(x1,...,xn) leq/eq 0 (i=1,...,m) xi >= 0 i=1,...,n (optional) Approaches: Classical: Differentiate the function and find points with a gradient of 0: problem: f has to be differentiable does not cope with constraints equation systems have to be solved that are frequently “nasty” (iterative algorithms such as Newton-Raphson’s method can be used). Lagrange multipliers are employed to cope with constraints. if g1,...,gm and f are linear: linear programming can be used. In the case that at least one function is non-linear general analytical solutions do no longer exist; and iteration algorithms have to be used.

2 General Teaching Plan EC for Numerical Optimization
Problem Definition (chapter 5 of textbook) Mutation and Crossover Operators for Real Numbers (extracts of chapter 6 and 7 of the textbook) The Evolution Strategy Approach (chapter 8 of textbook) The Penalty Function Approach and Other Approaches to Cope with Constraints Gray Codes [Practical Systems at Work]

3 Popular Numerical Methods
Newton-Raphson’s Method to solve: f(x)=0 f(x) is approximated by its tangent at the point (xn, f(xn)) and xn+1 is taken as the abcissa of the point of intersection of the tangent with the x-acis; that is, xn+1 is determined using: f(xn) + (xn+1xn)f’(xn) = 0 xn+1 = xn + hn with hn = (f(xn) / f’(xn)) the iterations are broken off when |hn| is less than the largest tolerable error. The Simplex Method is used to optimize a linear function with a set of linear constraints (linear programming). Quadratic programming [31] optimizes a quadratic function with linear constraints. Other interation methods (similar to Newton’s method) relying on xv+1= xv vdv where dv is a direction and v denotes the “jump” performed in the particular direction. Use quadratic/linear approximations of the optimization problem, and solve the optimization problem in the approximated space. Other popular optimization methods: the penalty trajectory method [220], the sequential quadratic penalty function method, and the SOLVER method [80].

4 Numerical Optimization with GAs
Coding alternatives include: binary coding Gray codes real-valued GAs Usually lower and upper bounds for variables have to be provided as the part of the optimization problem. Typical operators include: standard mutation and crossover non-uniform and boundary mutation arithmetical , simple, and heuristic crossover Constraints are a major challenge for function optimization. Ideas to cope with the problem include: elimination of equations through variable reduction. values in a solution are dynamic: they are nolonger independent of each other, but rather their contents is constrainted by the contents of other variables of the solution: in some cases a bound for possible changes can be computed (e.g. for convex search spaces (GENOCOP)). penalty functions. repair algorithms (GENETIC2)

5 Penalty Function Approach
Problem: f(x1,...,xn) has to be maximized with constraints gi(x1,...,xn) leq/eq 0 (i=1,...,m) define a new function: f’(x1,...,xn)= f(x1,...,xn) + i=1,mwihi (x1,...,xn) with: For gi(x1,...,xn) = 0: hi(x1,...,xn):= gi(x1,...,xn) For gi(x1,...,xn) <= 0: hi(x1,...,xn):= IF gi(x1,...,xn) < 0 THEN 0 ELSE gi(x1,...,xn) Remarks Penalty Function Approach: needs a lot of fine tuning, especially the selection of weights wi is very critical for the performance of the optimizer. frequently, the GA gets deceived only exploring the space of illegal solution, especially if penalties are too low; on the other hand, situations of premature convergence can arise when the GA terminates with a local minimum that is surrounded by illegal solutions, so that the GA cannot escape the local minimum, because the penalty for traversing illegal solutions is too high. a special approach called sequential quadratic penalty function method[9,39] has gained significant popularity.

6 Sequential Quadratic Penalty Function Method
Idea: instead of optimizing the constrainted function f(x), optimize: F(x,r) = f(x) + (1/(2r))(h1(x)2+...+hm(x)2) It has been shown by Fiacco et al. [189] that the solutions of optimizing the constrainted function f and the solutions of optimizing F are identical for r--0. However, it turned out to be difficult to minimize F in the limit with Newton’s method (see Murray [220]). More recently, Broyden and Attila [39,40] found a more efficient method; GENOCOP II that is discussed in our textbook employs this method.

7 Basic Loop of the SQPF Method
1) Differentiate F(x,r) yielding F’(x,r); 2) Choose a starting vector x0, choose a starting value ro>0; 3) r’:= ro; x’:=x0; REPEAT Solve F’(x,r’)=G(x)=0 for starting vector x’ yielding vector x1; x’:=x1; Decrease r’ by division through >1 UNTIL r’ is sufficiently close to 0; RETURN(x’);

8 Thoughts on Mutation Operators
Let t be the current generation# T be maximum generation number b be the degree of non-uniformity r be a random number in [0,1] Example mutation functions (return numbers in [0,y]): Mut1(t,y)=r*y Mut2(t,y)=y*(1-r(1-t/T)b ) Mut3(t,y)=y*r*(1-t/T)b

9 Various Numerical Crossover Operators
Let p1=(x1,y1) and p2=(x2,y2); crossover operators crossover(p1,p2) include: simple crossover: maxa(x1,y2a+y1 (1-a)); maxa(x2,y1a+y2(1-a)) arithmetical crossover: ap1 + (1-a)p2 with a[0,1] heuristic crossover(Wright[312]): p1 + (p1p2)a with a[0,1] if f(p1)>f(p2) Example: let p1=(1,2), p2=(5,1) be points a convex 2D-space: x2+y2 leq 28 and f(p1)>f(p2) a=1.0 phc=(-3,3) a=0.25 phc’=(0, 2.25) p1=(1,2) psc1=(5,1.7) p2=(5,1) psc2=(1,1) simple crossover yields: (1,1) and (5,sqrt(3)) (25+3=28). arithmetical crossover yields: all points along the line between p1 and p2. heuristic crossover yields: all points along the line between p1 and phc=(-3,3).

10 Another Example (Crossover Operators)
Let p1=(0,0,0) and p2=(1,1,1) in an unconstrainted search space: arithmetical crossover produces: (a,a,a) with a[0,1] simple crossover produces: (0,0,1), (0,1,1), (1,0,0), and (1,1,0). heuristic crossover produces: (a,a,a) with a[1,2], if f((1,1,1))>f((0,0,0)) (a,a,a) with a[-1,0], if f((1,1,1))<f((0,0,0)) (1,1,1) (0,0,0)

11 Problems of Optimization with Constraints
legal solutions illegal solutions illegal solutions S S S+ S S legal solutions S:= a solution S+:= the optimal solution

12 A Harder Optimization Problem
legal solutions legal solutions illegal solutions illegal solutions legal solutions

13 A Friedly Convex Search Space
illegal solutions pu p1 p legal solutions p2 illegal solutions pl illegal solutions Convexity (1) p1 and p2 in S => all points between p1 and p2 are in S (2) p in S => exactly two borderpoints can be found: pu and pl

14 Evolution Strategies Originally developed in Germany in the early 60s with Rechenberg and Schwefel being the main contributors. Main ideas include: floating point representation with standard derivation: (x,) where x is a vector in n-dimensional space and  is the standard derivation that influces how solutions are changed: x’= x + N(0, ). mutation is the sole operator. in some approaches  is changed dynamically (e.g. using Rechenberg’s 1/5 Rule). employs a two-membered evolution strategy. Was later generalized to support multimembered evolution strategies: employs: uniform crossover and averaging crossover. each member of the population has the same chance to reproduce (selection doesn’t consider fitness). weakest individual is eliminated to keep a constant population size.

15 -ES and -ESs.
developped by H.P. Schwefel: 2-member population ((1+1)-ES) is generalized to multimembered populations. 2 approaches are supported ()-ES:  indivividuals produceoffsprings the population consisting ofindividuals (the old generation and the parents) is reduced to  using selection. relies on replacement ()-ESs: lifetime of individuals is limited to one generation.  individual produce  offsprings ( > ) with the best  surviving. generates the new generation from the scratch Moreover, the standard deviation undergoes evolution.

16 Evolutionsstrategie and GAs
Differences ES and trad. GAs : real-code(ES) vs. binary string representation(GA) selection is performed implicitely by removing unfit individual deterministically (ES); GAs employ a stoachastic selection process, and does not rely on removal.. selection after recombination(ES); selection before recombination(GA). different handling of constraints: ES supports unequalities as a part of the problem specification, and disqualifies illegal offstring; moreover, ES adjusts control parameters if illegal offspring occur too frequently. GAs, on the other hand, employ penalty functions. mutation is less important for traditional GAs; crossover is less important for ESs. Some scientists, e.g. Fogel, claim that ES and GAs are not fundamentally different (see also Hoffmeister’s paper [141]).


Download ppt "Numerical Optimization"

Similar presentations


Ads by Google