CS5321 Numerical Optimization

Slides:



Advertisements
Similar presentations
Zhen Lu CPACT University of Newcastle MDC Technology Reduced Hessian Sequential Quadratic Programming(SQP)
Advertisements

C&O 355 Mathematical Programming Fall 2010 Lecture 6 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A.
Instabilities of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small.
Optimization.
Optimization Methods TexPoint fonts used in EMF.
Optimization of thermal processes
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
1 L-BFGS and Delayed Dynamical Systems Approach for Unconstrained Optimization Xiaohui XIE Supervisor: Dr. Hon Wah TAM.
Numerical Optimization
Tutorial 12 Unconstrained optimization Conjugate gradients.
1 L-BFGS and Delayed Dynamical Systems Approach for Unconstrained Optimization Xiaohui XIE Supervisor: Dr. Hon Wah TAM.
The Widrow-Hoff Algorithm (Primal Form) Repeat: Until convergence criterion satisfied return: Given a training set and learning rate Initial:  Minimize.
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Unconstrained Optimization Problem
An Introduction to Optimization Theory. Outline Introduction Unconstrained optimization problem Constrained optimization problem.

Computational Optimization
UNCONSTRAINED MULTIVARIABLE
Optimization Methods.
84 b Unidimensional Search Methods Most algorithms for unconstrained and constrained optimisation use an efficient unidimensional optimisation technique.
Optimization in Engineering Design Georgia Institute of Technology Systems Realization Laboratory 101 Quasi-Newton Methods.
Qualifier Exam in HPC February 10 th, Quasi-Newton methods Alexandru Cioaca.
Fin500J: Mathematical Foundations in Finance
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
EASTERN MEDITERRANEAN UNIVERSITY Department of Industrial Engineering Non linear Optimization Spring Instructor: Prof.Dr.Sahand Daneshvar Submited.
“On Sizing and Shifting The BFGS Update Within The Sized-Broyden Family of Secant Updates” Richard Tapia (Joint work with H. Yabe and H.J. Martinez) Rice.
Quasi-Newton Methods of Optimization Lecture 2. General Algorithm n A Baseline Scenario Algorithm U (Model algorithm for n- dimensional unconstrained.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
(COEN507) LECTURE III SLIDES By M. Abdullahi
Inequality Constraints Lecture 7. Inequality Contraints (I) n A Review of Lagrange Multipliers –As we discussed last time, the first order necessary conditions.
D Nagesh Kumar, IISc Water Resources Systems Planning and Management: M2L2 Introduction to Optimization (ii) Constrained and Unconstrained Optimization.
Optimal Control.
Function Optimization
Non-linear Minimization
CSLT ML Summer Seminar (13)
Does the set S SPAN R3 ?.
Computational Optimization
Quasi-Newton Methods Problem: SD, CG too slow to converge if NxN H matrix is ill-conditioned. SD: dx = - g (slow but no inverse to store or compute) QN:
Structure Refinement BCHM 5984 September 7, 2009.
Computational Optimization
Chap 9. General LP problems: Duality and Infeasibility
Ellipse Fitting COMP 4900C Winter 2008.
CS5321 Numerical Optimization
Chap 3. The simplex method
Collaborative Filtering Matrix Factorization Approach
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
CS5321 Numerical Optimization
CS5321 Numerical Optimization
L5 Optimal Design concepts pt A
Inverse & Identity MATRICES Last Updated: October 12, 2005.
Optimization Methods TexPoint fonts used in EMF.
CS5321 Numerical Optimization
CS5321 Numerical Optimization
I.4 Polyhedral Theory (NW)
Solving Linear Systems: Iterative Methods and Sparse Systems
I.4 Polyhedral Theory.
Linear Algebra review (optional)
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Outline Preface Fundamentals of Optimization
What are optimization methods?
Section 3: Second Order Methods
Numerical Analysis Lecture11.
CS5321 Numerical Optimization
Chapter 6. Large Scale Optimization
Conjugate Direction Methods
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Presentation transcript:

CS5321 Numerical Optimization 06 Quasi-Newton Methods 11/24/2018

Hessian matrix The Hessian matrix B is needed in computing Newton’s direction pk = −Bk−1fk. What needed is the inverse multiplying a vector. Approximate result maybe good enough in use. Quasi-Newton: use gradient vector to approximate Hessian DFP and BFGS updating formula SR1 (Symmetric Rank 1 update) method Broyden class 11/24/2018

The secant formula Suppose we have Hessian Bk at current point xk Next point is The model at xk+1 is Assume and let and Then we have the secant formula Also require Bk+1 to be spd x k + 1 = ® p m k + 1 ( p ) = f g T 2 B r m k + 1 ( ¡ ® p ) = g B s k = x + 1 ¡ ® p y k = g + 1 ¡ B k + 1 s = y s T k B + 1 = y > 11/24/2018

DFP updating formula The problem of approximating Bk+1 becomes subject to With some special norm, the solution is This is the DFP updating formula Davidon, Fletcher and Powell m i n B k ¡ B = T ; s k y B k + 1 = ( I ¡ ½ y s T ) ½ k = 1 y T s 11/24/2018

BFGS updating formula Let . Similar results can be obtained subject to The solution is This is the BFGS updating formula Broyden, Fletcher, Goldfarb, and Shanno H k = B ¡ 1 m i n H k ¡ H = T ; y k s H k + 1 = ( I ¡ ½ s y T ) ½ k = 1 y T s 11/24/2018

More on BFGS The initial H0 need be constructed Finite difference or automatic differentiation (chap 8) Using identity matrix or Hessian Bk can be constructed via the Sherman-Morrison-Woodbury formula (rank 2 update) If Hk is positive definite, Hk+1 is positive definite. ( y T k s = ) B k + 1 = ¡ s T y 11/24/2018

The SR1 method Desire a symmetric rank 1 update, Bk+1=Bk+vvT, that satisfies the secant constrain: yk=Bk+1sk. Vector v is parallel to (yk−Bksk). Let v = (yk−Bksk) Substitute back to the secant constrain Let zk = yk−Bksk. The SR1 is y k = ( B + ½ v T ) s ½ = s i g n [ T k ( y ¡ B ) ] ; ± § 1 2 B k + 1 = z T s 11/24/2018

More on the SR1 Let wk = sk− Hkyk. By Sherman-Morrison formula, If yk= Bksk, Bk+1= Bk. If yk Bksk but (yk−Bksk)Tsk=0, there is no the SR1. If (yk− Bksk)Tsk  0, the SR1 is numerical instable. Use Bk+1= Bk. In practice, the Bk generated by the SR1 satisfies H k + 1 = w T y l i m k ! 1 B ¡ r 2 f ( x ¤ ) = 11/24/2018

The Broyden class The Broyden class is a family of quasi-Newton updating formulas that have the form k is a scalar function. It is a linear combination of BFGS and DFP It is a restricted Broyden class if [0,1] B k + 1 = ¡ s T y Á ( ) v v k = µ y T s ¡ B ¶ B r o y d e n k + 1 = ( ¡ Á ) F G S D P 11/24/2018

More on the Broyden class The SR1 is also belong to the Broyden class with But may not be belong to the restricted Broyden class. If k+1  0 and Bk is positive definite, then Bk is positive definite. Á k = s T y ¡ B 11/24/2018

Convergence of BFGS Global convergence Convergent rate For any spd B0 and any x0, if f is twice continuously differentiable and the working set is convex with properly bounded Hessian, then the BFGS converges to the minimizer x* of f. Convergent rate If f is twice continuously differentiable and 2f is Lipschitz continuous near x*, and then the BFGS converges superlinearly 1 X k = x ¡ ¤ < 11/24/2018