CS5321 Numerical Optimization

Slides:



Advertisements
Similar presentations
Optimization.
Advertisements

Optimization of thermal processes
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Steepest Decent and Conjugate Gradients (CG). Solving of the linear equation system.
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
Cojugate Gradient Method Zhengru Zhang ( 张争茹 ) Office: Math. Building 413(West) 2010 年教学实践周
2.7.6 Conjugate Gradient Method for a Sparse System Shi & Bo.
Jonathan Richard Shewchuk Reading Group Presention By David Cline
1cs542g-term Notes  Assignment 1 due tonight ( me by tomorrow morning)
Numerical Optimization
Function Optimization Newton’s Method. Conjugate Gradients
1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me .
Tutorial 12 Unconstrained optimization Conjugate gradients.
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Function Optimization. Newton’s Method Conjugate Gradients Method
Advanced Topics in Optimization
Monica Garika Chandana Guduru. METHODS TO SOLVE LINEAR SYSTEMS Direct methods Gaussian elimination method LU method for factorization Simplex method of.
CSE 245: Computer Aided Circuit Simulation and Verification
Why Function Optimization ?

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
Introduction to Numerical Analysis I MATH/CMPSC 455 Conjugate Gradient Methods.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Example: Introduction to Krylov Subspace Methods DEF: Krylov sequence
Numerical Analysis 1 EE, NCKU Tien-Hao Chang (Darby Chang)
UNCONSTRAINED MULTIVARIABLE
84 b Unidimensional Search Methods Most algorithms for unconstrained and constrained optimisation use an efficient unidimensional optimisation technique.
Qualifier Exam in HPC February 10 th, Quasi-Newton methods Alexandru Cioaca.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Incorporating Iterative Refinement with Sparse Cholesky April 2007 Doron Pearl.
CSE 245: Computer Aided Circuit Simulation and Verification Matrix Computations: Iterative Methods I Chung-Kuan Cheng.
Case Study in Computational Science & Engineering - Lecture 5 1 Iterative Solution of Linear Systems Jacobi Method while not converged do { }
Chapter 10 Minimization or Maximization of Functions.
1 Chapter 6 General Strategy for Gradient methods (1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Steepest Descent.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Krylov-Subspace Methods - I Lecture 6 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Network Systems Lab. Korea Advanced Institute of Science and Technology No.1 Maximum Norms & Nonnegative Matrices  Weighted maximum norm e.g.) x1x1 x2x2.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
The Landscape of Sparse Ax=b Solvers Direct A = LU Iterative y’ = Ay Non- symmetric Symmetric positive definite More RobustLess Storage More Robust More.
Krylov-Subspace Methods - I
Solving Poisson Equations Using Least Square Technique in Image Editing Colin Zheng Yi Li.
CSE 245: Computer Aided Circuit Simulation and Verification
CSE 245: Computer Aided Circuit Simulation and Verification
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Non-linear Least-Squares
Conjugate Gradient Problem: SD too slow to converge if NxN H matrix is ill-conditioned. SD: dx = - g (slow but no inverse to store or compute) CG: dx =
Outline Single neuron case: Nonlinear error correcting learning
CSE 245: Computer Aided Circuit Simulation and Verification
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Conjugate Gradient Method
Ch2: Adaline and Madaline
Introduction to Scientific Computing II
CS5321 Numerical Optimization
~ Least Squares example
~ Least Squares example
Administrivia: November 9, 2009
Performance Optimization
Outline Preface Fundamentals of Optimization
L23 Numerical Methods part 3
Outline Preface Fundamentals of Optimization
Section 3: Second Order Methods
CS5321 Numerical Optimization
Conjugate Gradient Optimization
Conjugate Direction Methods
CS5321 Numerical Optimization
CS5321 Numerical Optimization
First-Order Methods.
Presentation transcript:

CS5321 Numerical Optimization 05 Conjugate Gradient Methods 12/30/2018

Conjugate gradient methods For convex quadratic problems, the steepest descent method is slow in convergence. the Newton’s method is expensive in solving Ax=b. the conjugate gradient method solves Ax=b iteratively. Outline Conjugate directions Linear conjugate gradient method Nonlinear conjugate gradient method 12/30/2018

Quadratic optimization problem Consider the quadratic optimization problem A is symmetric positive definite The optimal solution is at f (x) = 0 Define r(x) = f (x) = Ax − b (the residual). Solve Ax=b without inverting A. (Iterative method) m i n f ( x ) = 1 2 T A ¡ b r ( 1 2 x T A ¡ b ) = 12/30/2018

Steepest descent+line search Given an initial guess x0. The search direction: pk= −fk= −rk= b −Axk The optimal step length: The optimal solution is Update xk+1= xk + kpk. Goto 2. m i n f ( x ) = 1 2 T A ¡ b m i n ® f ( x k + p ) ® k = ¡ r T p A 12/30/2018

Conjugate direction For a symmetric positive definite matrix A, one can define A-inner-product as x, yA= xTAy. A-norm is defined as Two vectors x and y are A-conjugate for a symmetric positive definite matrix A if xTAy=0 x and y are orthogonal under A-inner-product. The conjugate directions are a set of search directions {p0, p1, p2,…}, such that piApj=0 for any i  j. k x A = p T 12/30/2018

Example A = µ 1 4 ¶ ; b x Conjugate direction Steepest descent ¶ ; b x Conjugate direction Steepest descent 12/30/2018

Conjugate gradient A better result can be obtained if the current search direction combines the previous one pk+1= −rk + kpk Let pk+1 be A-conjugate to pk. ( ) p T k + 1 A = p T k + 1 A = ¡ r ¯ ¯ k = p T A r + 1 12/30/2018

The linear CG algorithm With some linear algebra, the algorithm can be simplified as Given x0, r0=Ax0−b, p0= −r0 For k = 0,1,2,… until ||rk|| = 0 ® k = r T p A x + 1 ¯ ¡ 12/30/2018

Properties of linear CG One matrix-vector multiplication per iteration. Only four vectors are required. (xk, rk, pk, Apk) Matrix A can be stored implicitly The CG guarantees convergence in r iterations, where r is the number of distinct eigenvalues of A If A has eigenvalues 1  2  …  n, k x + 1 ¡ ¤ 2 A · µ ¸ n ¶ 12/30/2018

CG for nonlinear optimization The Fletcher-Reeves method Given x0. Set p0= −f0, For k = 0,1,…, until f0=0 Compute optimal step length k and set xk+1=xk+ kpk Evaluate fk+1 ¯ k + 1 = r f T p ¡ 12/30/2018

Other choices of  Polak-Ribiere: Hestens-Siefel: Y.Dai and Y.Yuan ¯ k + 1 = r f T ( ¡ ) Polak-Ribiere: Hestens-Siefel: Y.Dai and Y.Yuan (1999) WW.Hager and H.Zhang (2005) ¯ k + 1 = r f T ( ¡ ) p ¯ k + 1 = r f 2 ( ¡ ) T p ¯ k + 1 = µ y ¡ 2 p T ¶ r f 12/30/2018