Instructor: Prof.Dr.Sahand Daneshvar Presented by: Seyed Iman Taheri Student number: 136023 Non linear Optimization Spring 2014-15 EASTERN MEDITERRANEAN.

Slides:



Advertisements
Similar presentations
Boyce/DiPrima 9th ed, Ch 2.4: Differences Between Linear and Nonlinear Equations Elementary Differential Equations and Boundary Value Problems, 9th edition,
Advertisements

Ch 3.2: Solutions of Linear Homogeneous Equations; Wronskian
Interpolation A standard idea in interpolation now is to find a polynomial pn(x) of degree n (or less) that assumes the given values; thus (1) We call.
Mathematics1 Mathematics 1 Applied Informatics Štefan BEREŽNÝ.
CSE 330: Numerical Methods
Copyright © Cengage Learning. All rights reserved.
Engineering Mathematics Class #15 Fourier Series, Integrals, and Transforms (Part 3) Sheng-Fang Huang.
Optimization of thermal processes
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Ch 5.8: Bessel’s Equation Bessel Equation of order :
Thursday, April 25 Nonlinear Programming Theory Separable programming Handouts: Lecture Notes.
Chapter 7 Numerical Differentiation and Integration
Nonlinear programming: One dimensional minimization methods
MIT and James Orlin © Nonlinear Programming Theory.
PHYS2020 NUMERICAL ALGORITHM NOTES ROOTS OF EQUATIONS.
Function Optimization Newton’s Method. Conjugate Gradients
Page 1 Page 1 Engineering Optimization Second Edition Authors: A. Rabindran, K. M. Ragsdell, and G. V. Reklaitis Chapter-2 (Functions of a Single Variable)
Tutorial 12 Unconstrained optimization Conjugate gradients.
Optimization Methods One-Dimensional Unconstrained Optimization
Ch 6.1: Definition of Laplace Transform Many practical engineering problems involve mechanical or electrical systems acted upon by discontinuous or impulsive.
Ch 5.1: Review of Power Series
Ch 5.1: Review of Power Series Finding the general solution of a linear differential equation depends on determining a fundamental set of solutions of.
Ch 5.3: Series Solutions Near an Ordinary Point, Part II
Ch 3.3: Linear Independence and the Wronskian
Dr. Marco A. Arocha Aug,  “Roots” problems occur when some function f can be written in terms of one or more dependent variables x, where the.
NUMERICAL METHODS WITH C++ PROGRAMMING
12 INFINITE SEQUENCES AND SERIES The Comparison Tests In this section, we will learn: How to find the value of a series by comparing it with a known.
Optimization Methods One-Dimensional Unconstrained Optimization
Name: Mehrab Khazraei(145061) Title: Penalty or Exterior penalty function method professor Name: Sahand Daneshvar.
Chapter 3 Root Finding.
1 1 Slide © 2000 South-Western College Publishing/ITP Slides Prepared by JOHN LOUCKS.
Theorems on continuous functions. Weierstrass’ theorem Let f(x) be a continuous function over a closed bounded interval [a,b] Then f(x) has at least one.
Computational Optimization
f /(x) = 6x2 – 6x – 12 = 6(x2 – x – 2) = 6(x-2)(x+1).
19.5 Numeric Integration and Differentiation
Solving Non-Linear Equations (Root Finding)
Ch 5.3: Series Solutions Near an Ordinary Point, Part II A function p is analytic at x 0 if it has a Taylor series expansion that converges to p in some.
The importance of sequences and infinite series in calculus stems from Newton’s idea of representing functions as sums of infinite series.  For instance,
Nonlinear programming Unconstrained optimization techniques.
Copyright © 2013, 2009, 2005 Pearson Education, Inc. 1 3 Polynomial and Rational Functions Copyright © 2013, 2009, 2005 Pearson Education, Inc.
Lecture 3 Numerical Analysis. Solution of Non-Linear Equations Chapter 2.
Boyce/DiPrima 9 th ed, Ch 5.1: Review of Power Series Elementary Differential Equations and Boundary Value Problems, 9 th edition, by William E. Boyce.
Lecture 6 Numerical Analysis. Solution of Non-Linear Equations Chapter 2.
Method of Hooke and Jeeves
1 Solution of Nonlinear Equation Dr. Asaf Varol
Chapter 3 Roots of Equations. Objectives Understanding what roots problems are and where they occur in engineering and science Knowing how to determine.
Numerical Methods for Engineering MECN 3500
EASTERN MEDITERRANEAN UNIVERSITY Department of Industrial Engineering Non linear Optimization Spring Instructor: Prof.Dr.Sahand Daneshvar Submited.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
D Nagesh Kumar, IIScOptimization Methods: M2L4 1 Optimization using Calculus Optimization of Functions of Multiple Variables subject to Equality Constraints.
Boundary Value Problems l Up to this point we have solved differential equations that have all of their initial conditions specified. l There is another.
Numerical Methods.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Lecture 22 Numerical Analysis. Chapter 5 Interpolation.
12 INFINITE SEQUENCES AND SERIES. In general, it is difficult to find the exact sum of a series.  We were able to accomplish this for geometric series.
Numerical Methods Solution of Equation.
Optimization Theory O Professor: Dr. Sahand Daneshvar O Student’s Name: Milad Kermani (125512) M.S student of Mechanical Engineering Department Eastern.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Math 3120 Differential Equations with Boundary Value Problems
Boyce/DiPrima 9 th ed, Ch 6.1: Definition of Laplace Transform Elementary Differential Equations and Boundary Value Problems, 9 th edition, by William.
Boyce/DiPrima 9 th ed, Ch 5.3: Series Solutions Near an Ordinary Point, Part II Elementary Differential Equations and Boundary Value Problems, 9 th edition,
Ch 9.6: Liapunov’s Second Method In Section 9.3 we showed how the stability of a critical point of an almost linear system can usually be determined from.
CSE 330: Numerical Methods. What is true error? True error is the difference between the true value (also called the exact value) and the approximate.
Answers for Review Questions for Lectures 1-4. Review Lectures 1-4 Problems Question 2. Derive a closed form for the estimate of the solution of the equation.
11.4 The Comparison Tests In this section, we will learn:
SOLUTION OF NONLINEAR EQUATIONS
16. Mean Square Estimation
Presentation transcript:

Instructor: Prof.Dr.Sahand Daneshvar Presented by: Seyed Iman Taheri Student number: Non linear Optimization Spring EASTERN MEDITERRANEAN UNIVERSITY Department of Industrial Engineering

Line Search Using Derivatives bisection search method Newton's method

bisection search method Suppose that we wish to minimize a function θ over a closed and bounded interval. Furthermore, suppose that θ is pseudoconvex and hence, differentiable. At iteration k, let the interval of uncertainty be [a k, b k ]. Suppose that the derivative θ'(λ k ) is known, and consider the following three possible cases:

1. If θ'(λ k ) = 0, then, by the pseudoconvexity of θ, λ k is a minimizing point. 2. If θ'(λ k ) > 0, then, for λ > λ k we have θ'(λ k ) (λ- λ k ) > 0; and by the pseudoconvexity of θ it follows that θ(λ) ≥θ(λ k ). In other words, the minimum occurs to the left of λ k, so that the new interval of uncertainty [a k+l, b k+1 ]is given by [a k, λ k ]. 3. If θ'(λ k ) 0, so that θ(λ) ≥θ(λ k ). Thus, the minimum occurs to the right of λ k, so that the new interval of uncertainty [a k+l, b k+1 ] is given by [λ k, b k ].

The position of λ k, in the interval [a k, b k ] must be chosen so that the maximum possible length of the new interval of uncertainty is minimized. That is, λ k must be chosen so as to minimize the maximum of λ k - a k and b k - λ k. Obviously, the optimal location of λ k is the midpoint (1/2)(a k + b k ). To summarize, at any iteration k, θ' is evaluated at the midpoint of the interval of uncertainty. Based on the value of θ', we either stop or construct a new interval of uncertainty whose length is half that at the previous iteration. Note that this procedure is very similar to the dichotomous search method except that at each iteration, only one derivative evaluation is required, as opposed to two functional evaluations for the dichotomous search method. However, the latter is akin to a finite difference derivative evaluation.

Convergence of the Bisection Search Method Note that the length of the interval of uncertainty after n observations is equal to (1/2) n (b 1 -a 1 ), so that the method converges to a minimum point within any desired degree of accuracy. In particular, if the length of the final interval of uncertainty is fixed at C, then n must be chosen to be the smallest integer such that (1/2) n ≤ l /(b 1 -a 1 ).

Summary of the Bisection Search Method We now summarize the bisection search procedure for minimizing a pseudoconvex function θ over a closed and bounded interval. Initialization Step Let [a 1, b 1 ] be the initial interval of uncertainty, and let P be the allowable final interval of uncertainty. Let n be the smallest positive integer such that (1/2) n ≤ l /(b 1 -a 1 ). Let k = 1 and go to the Main Step. Main Step Let λ k = (1/2)(a k + b k ) and evaluate θ'(λ k ). If θ'(λ k )=0, stop; λ k is an optimal solution. Otherwise, go to Step 2 if θ'(λ k )> 0, and go to Step 3 if θ'(λ k )< Let a k+1 = a k and b k+1 = λ k. GO to Step Let a k+1 = λ k, and b k+1 = b k. GO to Step If k = n, stop; the minimum lies in the interval [ a n+1, b n+1 ]. Otherwise, replace k by k + 1 and repeat Step 1.

Example Consider the following problem: Minimize (λ) 2 + 2λ subject to -3≤λ≤6. Suppose that we want to reduce the interval of uncertainty to an interval whose length e is less than or equal to 0.2. Hence, the number of observations n satisfying (1/2) n ≤ l /(b 1 -a 1 )= 0.2/9 = is given by n=6. A summary of the computations using the bisection search method is given in Table of next slide.

Summary of Computations for the Bisection Search Method Note that the final interval of uncertainty is [ , ], so that the minimum could be taken as the midpoint,

Newton’s Method Newton’s method is based on exploiting the quadratic approximation of the function θ at a given point λ k. This quadratic approximation q is given by q(λ) = θ(λ k ) + θ΄ (λ k )(λ - λ k ) + (1/2)θ"̋(λ k )(λ - λ k ) 2. The point λk+1, is taken to be the point where the derivative of q is equal to zero. This yields θ΄ (λ k )+ θ"̋(λ k )(λ k+1 - λ k ) = 0, so that λ k+1 = λ k –(θ΄ (λ k )/θ"̋(λ k )) The procedure is terminated when |λk+1 - λk| < ε, or when |θ΄ (λ k )|< ε, where ε is a pre-specified termination scalar. Note that the above procedure can only be applied for twice differentiable functions. Furthermore, the procedure is well defined only if θ"̋(λ k )≠0 for each k.

Example Consider the function θ: Note that θ is twice differentiable everywhere. We apply Newton’s method, starting from two different points. In the first case, λ 1 = 0.40; and as shown in first Table of next slide, the procedure produces the point after six iterations. The reader can verify that the procedure indeed converges to the stationary point λ =0. In the second case, λ= 0.60, and the procedure oscillates between the points 0.60 and -0.60, as shown in second Table of next slide.

Summary of Computations for Newton's Method Starting from λ1 = 0.4 Summary of Computations for Newton‘s Method Starting from λ 1 = 0.6

Convergence of Newton's Method The method of Newton, in general, does not converge to a stationary point starting with an arbitrary initial point. Observe that, in general, Theorem cannot be applied as a result of the unavailability of a descent function. However, as shown in next Theorem, if the starting point is sufficiently close to a stationary point, then a suitable descent function can be devised so that the method converges.

Theorem

Proof

Thanks for your attention