CS5321 Numerical Optimization

Slides:



Advertisements
Similar presentations
Zhen Lu CPACT University of Newcastle MDC Technology Reduced Hessian Sequential Quadratic Programming(SQP)
Advertisements

Instabilities of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small.
Numerical Solution of Nonlinear Equations
Inversion Transforming the apparent to « real » resistivity. Find a numerical model that explains the field measurment.
Siddharth Choudhary.  Refines a visual reconstruction to produce jointly optimal 3D structure and viewing parameters  ‘bundle’ refers to the bundle.
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
1cs542g-term Notes  Extra class this Friday 1-2pm  Assignment 2 is out  Error in last lecture: quasi-Newton methods based on the secant condition:
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
Methods For Nonlinear Least-Square Problems
1 L-BFGS and Delayed Dynamical Systems Approach for Unconstrained Optimization Xiaohui XIE Supervisor: Dr. Hon Wah TAM.
Martin Burger Institut für Numerische und Angewandte Mathematik European Institute for Molecular Imaging CeNoS Total Variation and related Methods Numerical.
The Widrow-Hoff Algorithm (Primal Form) Repeat: Until convergence criterion satisfied return: Given a training set and learning rate Initial:  Minimize.
The Perceptron Algorithm (Dual Form) Given a linearly separable training setand Repeat: until no mistakes made within the for loop return:
Unconstrained Optimization Problem
Ordinary least squares regression (OLS)
12 1 Variations on Backpropagation Variations Heuristic Modifications –Momentum –Variable Learning Rate Standard Numerical Optimization –Conjugate.
Advanced Topics in Optimization
Computational Optimization
UNCONSTRAINED MULTIVARIABLE
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 9. Optimization problems.
ITERATIVE TECHNIQUES FOR SOLVING NON-LINEAR SYSTEMS (AND LINEAR SYSTEMS)
ENCI 303 Lecture PS-19 Optimization 2
Ordinary Least-Squares Emmanuel Iarussi Inria. Many graphics problems can be seen as finding the best set of parameters for a model, given some data Surface.
Qualifier Exam in HPC February 10 th, Quasi-Newton methods Alexandru Cioaca.
Application of Differential Applied Optimization Problems.
Nonlinear least squares Given m data points (t i, y i ) i=1,2,…m, we wish to find a vector x of n parameters that gives a best fit in the least squares.
Fin500J: Mathematical Foundations in Finance
Section 4-1: Introduction to Linear Systems. To understand and solve linear systems.
LTSI (1) Faculty of Mech. & Elec. Engineering, University AL-Baath, Syria Ahmad Karfoul (1), Julie Coloigner (2,3), Laurent Albera (2,3), Pierre Comon.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
MAT 1228 Series and Differential Equations Section 3.7 Nonlinear Equations
A comparison between PROC NLP and PROC OPTMODEL Optimization Algorithm Chin Hwa Tan December 3, 2008.
Circuits Theory Examples Newton-Raphson Method. Formula for one-dimensional case: Series of successive solutions: If the iteration process is converged,
linear  2.3 Newton’s Method ( Newton-Raphson Method ) 1/12 Chapter 2 Solutions of Equations in One Variable – Newton’s Method Idea: Linearize a nonlinear.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
1 Chapter 6 General Strategy for Gradient methods (1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Steepest Descent.
Variations on Backpropagation.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Differential Equations Linear Equations with Variable Coefficients.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Optimal Control.
Bounded Nonlinear Optimization to Fit a Model of Acoustic Foams
CSCE 441: Computer Graphics Forward/Inverse kinematics
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Non-linear Least-Squares
Some useful linear algebra
Collaborative Filtering Matrix Factorization Approach
CS5321 Numerical Optimization
Digital Visual Effects Yung-Yu Chuang
Variations on Backpropagation.
Outline Single neuron case: Nonlinear error correcting learning
CS5321 Numerical Optimization
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
CS5321 Numerical Optimization
Sec 21: Analysis of the Euler Method
Optimization Methods TexPoint fonts used in EMF.
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Variations on Backpropagation.
CS5321 Numerical Optimization
Numerical Analysis – Solving Nonlinear Equations
Computer Animation Algorithms and Techniques
CS5321 Numerical Optimization
Multiple linear regression
CS5321 Numerical Optimization
Presentation transcript:

CS5321 Numerical Optimization 10 Least Squares Problem 5/9/2019

Least-squares problems Linear least-squares problems QR method Nonlinear least-squares problems Gradient and Hessian of nonlinear LSP Gauss--Newton method Levenberg--Marquardt method Methods for large residual problem 5/9/2019

Example of linear least square y = β1 + β2x (from Wikipedia) β1 = 3.5, β2 = 1.4. The line is y = 3.5 + 1.4x 5/9/2019

Linear least squares problems A linear least-squares problem is f (x)=1/2||Ax−y||2. It’s gradient is f (x)=AT(Ax−y) The optimal solution is at f (x)=0, ATAx=ATy ATAx = ATy is called the normal equation. Perform QR decomposition on matrix A = QR. RT is invertible. The solution x = R−1QTy. A T x = R Q y 5/9/2019

Example of nonlinear LS Á ( x ; t ) = 1 + 2 3 4 e ¡ 5 Find (x1,x2,x3,x4,x5) to minimize 1 2 m X j = [ Á ( x ; t ) ¡ y ] 5/9/2019

Gradient and Hessian of LSP The object function of least squares problem is where ri are n variable functions. Define The Jacobian Gradient Hessian f ( x ) = 1 2 m X j r R ( x ) = B @ r 1 2 . m C A J ( x ) = B @ r 1 T 2 . m C A r f ( x ) = m X j 1 J T R 2 + 5/9/2019

Gauss-Newton method Gauss-Newton uses the Hessian approximation 2 f ( x ) = J T + m X j 1 Gauss-Newton uses the Hessian approximation It’s a good approximation if ||R|| is small. This is the matrix of the normal equation Usually with the line search technique Replace with r 2 f ( x ) ¼ J T f ( x ) = 1 2 m X j r f ( x ) = 1 2 k J p + R 5/9/2019

Convergence of Gauss-Newton Suppose each rj is Lipschitz continuously differentiable in a neighborhood N of {x|f(x)f(x0)} and the Jacobians J(x) satisfy ||J(x)z||||z||. Then the Gauss-Newton method, withk that satisfies the Wolfe conditions, has l i m k ! 1 J T R = 5/9/2019

Levenberg-Marquardt method Gauss-Newton + trust region The problem becomes subject to || p ||  k Optimal condition: (recall that in chap 4) Equivalent linear least-square problem m i n p 1 2 k J + R ( J T + ¸ I ) p = ¡ R ¢ k m i n p 1 2 ° µ J ¸ I ¶ + R 5/9/2019

Convergence of Levenberg-Marquardt Suppose L={x | f (x)  f (x0)} is bounded and each rj is Lipschitz continuously differentiable in a neighborhood N of L. Assume for each k, the approximation solution pk of the Levenberg-Marquardt method satisfies the inequality for some constant c1>0, and ||pk||k for some >1. Then m k ( ) ¡ p ¸ c 1 J T r i n µ ¢ ; ¶ l i m k ! 1 J T R = 5/9/2019

Large residual problem 2 f ( x ) = J T + m X j 1 When the second term of the Hessian is large Use quasi-Newton to approximate the second term The secant equation of 2rj(x) is The secant equation of the second term and the update formula (next slide) ( B j ) x k + 1 ¡ = r 5/9/2019

Dennis, Gay, Welsch update formula. k + 1 ( x ¡ ) = m X j r B [ ] J T R Dennis, Gay, Welsch update formula. S k + 1 = ( z ¡ s ) y T 2 s = x k + 1 ¡ y J T r z 5/9/2019