Curved Trajectories towards Local Minimum of a Function Al Jimenez Mathematics Department California Polytechnic State University San Luis Obispo, CA 93407.

Slides:



Advertisements
Similar presentations
Zhen Lu CPACT University of Newcastle MDC Technology Reduced Hessian Sequential Quadratic Programming(SQP)
Advertisements

Instabilities of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small.
13-Optimization Assoc.Prof.Dr. Ahmet Zafer Şenalp Mechanical Engineering Department Gebze Technical.
Optimization with Constraints
Optimization.
Engineering Optimization
Optimization : The min and max of a function
Introducción a la Optimización de procesos químicos. Curso 2005/2006 BASIC CONCEPTS IN OPTIMIZATION: PART II: Continuous & Unconstrained Important concepts.
Optimization of thermal processes
Optimization 吳育德.
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
The loss function, the normal equation,
Function Optimization Newton’s Method. Conjugate Gradients
1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me .
Optimization Methods One-Dimensional Unconstrained Optimization
Gradient Methods May Preview Background Steepest Descent Conjugate Gradient.
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Unconstrained Optimization Problem
Gradient Methods Yaron Lipman May Preview Background Steepest Descent Conjugate Gradient.
Improved BP algorithms ( first order gradient method) 1.BP with momentum 2.Delta- bar- delta 3.Decoupled momentum 4.RProp 5.Adaptive BP 6.Trinary BP 7.BP.
Function Optimization. Newton’s Method Conjugate Gradients Method
Advanced Topics in Optimization
Why Function Optimization ?
Math for CSLecture 51 Function Optimization. Math for CSLecture 52 There are three main reasons why most problems in robotics, vision, and arguably every.
Optimization Methods One-Dimensional Unconstrained Optimization
Tier I: Mathematical Methods of Optimization

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
UNCONSTRAINED MULTIVARIABLE
Collaborative Filtering Matrix Factorization Approach
Geometry Optimisation Modelling OH + C 2 H 4 *CH 2 -CH 2 -OH CH 3 -CH 2 -O* 3D PES.
ENCI 303 Lecture PS-19 Optimization 2
Nonlinear programming Unconstrained optimization techniques.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Multivariate Unconstrained Optimisation First we consider algorithms for functions for which derivatives are not available. Could try to extend direct.
Engineering Optimization Chapter 3 : Functions of Several Variables (Part 1) Presented by: Rajesh Roy Networks Research Lab, University of California,
559 Fish 559; Lecture 5 Non-linear Minimization. 559 Introduction Non-linear minimization (or optimization) is the numerical technique that is used by.
Quasi-Newton Methods of Optimization Lecture 2. General Algorithm n A Baseline Scenario Algorithm U (Model algorithm for n- dimensional unconstrained.
Dan Simon Cleveland State University Jang, Sun, and Mizutani Neuro-Fuzzy and Soft Computing Chapter 6 Derivative-Based Optimization 1.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Bundle Adjustment A Modern Synthesis Bill Triggs, Philip McLauchlan, Richard Hartley and Andrew Fitzgibbon Presentation by Marios Xanthidis 5 th of No.
Chapter 10 Minimization or Maximization of Functions.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
1 Chapter 6 General Strategy for Gradient methods (1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Steepest Descent.
Steepest Descent Method Contours are shown below.
Gradient Methods In Optimization
Survey of unconstrained optimization gradient based algorithms
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
METHOD OF STEEPEST DESCENT ELE Adaptive Signal Processing1 Week 5.
Data assimilation for weather forecasting G.W. Inverarity 06/05/15.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Krylov-Subspace Methods - II Lecture 7 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Non-linear Minimization
Computational Optimization
Iterative Non-Linear Optimization Methods
CS5321 Numerical Optimization
Non-linear Least-Squares
Collaborative Filtering Matrix Factorization Approach
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
Optimization Part II G.Anuradha.
Instructor :Dr. Aamer Iqbal Bhatti
~ Least Squares example
The loss function, the normal equation,
Mathematical Foundations of BME Reza Shadmehr
~ Least Squares example
Performance Optimization
Outline Preface Fundamentals of Optimization
Section 3: Second Order Methods
Presentation transcript:

Curved Trajectories towards Local Minimum of a Function Al Jimenez Mathematics Department California Polytechnic State University San Luis Obispo, CA …Taylor Series and Rotations Spring, 2008

Introduction and Notation The Problem Derivatives: A local min x * is a critical point: Necessary condition: 0

Typical Iterative Methods Sequence is generated from x 0 Such that With v k a vector with property a descent direction And p k > 0 typically approximates solution of called the line search or the scalar search Proven to converge for smooth functions

Current Methods Selecting v k has huge effect on convergence rate: –Steepest Descent: 1 st order –Newtons direction: 2 nd order, but may not be a descent direction when far from a min –Conjugate Directions uses v k-1, v k-2,... –Quasi-Newton/Variable metric also uses v k-1, v k-2,... –High order Tensor models fit prior iteration values –Number of derivatives available affects method The scalar search –Accuracy of scalar minimization –Quadratic models: Trust Region

Infinite Series of Solution Matrix vector products, but shown with exponents for connections with scalar Taylor series.

Infinite Series of Solution… Define: Then: For p = 1:

Curved Trajectories Algorithm At k th iteration, estimate, then calculate: Select order, modify d i, and select p k 2 nd order: 3 rd order: 4 th order:

Challenges High order terms accurately approximated from the Gradient and the Hessian Scalar searches along polynomial curved trajectories Performance for large problems –Exploit Sparse Hessian Store nonzeros only, no operations on zeros Far from solution: –Hessian not positive definite (solved) Hessian modified and use CG step as last resort

Hessian < 0 Changes

Cuter Performance Profiles

Current Research Pursuits Handle multiple functions: Pareto optimal points Handle Constraint Functions Explore the family of infinite series for combination of composition functions.

Rosenbrock Banana Function Algorithm selects

x0x0 x1x1 x2x2 x3x3 f = 24.2 f = 4 f = 0.5

3D View

Trajectories from starting point

Rotations

Rotations 3D

Rotations At point we have h ( p ) is trajectory and R (θ) is rotation matrix. h (0) = 0 and R (0) = I, and for 2 coordinates, counterclockwise At the k th step far from solution we want: But settle for p k, θ k :

Rotations (continued) Gives Trajectory angle with the gradient for R (0) = I Observations:

Rotation Challenges/Results Select effective θ k without too much work –Using existing strategy to calculate p k, then calculate a θ k from θ * and θ G. Then calculate a new p k again using rotated trajectory. –Good results with –θ k > 40º indicates elongated ellipse contours, and rotation seems unproductive in this case. –Effective when CTA series is convergent and iteration is not close to the minimum point. Functions of more than 2 variables later

f (p, θ)

f (p, θ), θ = 0, -0.1, -0.2, -0.3 θ = 0 θ = -0.1 θ = -0.2 θ = -0.3

More than Two Coordinates Ignore coordinates with insignificant Newton correction magnitudes. Success achieved by adding the 3rd coordinate to the first two as follows: –Calculate the rotation by paring the 3 rd coordinate with each of the top 2 coordinates. –This results in a rotation matrix: –Where the angles θ 1, θ 2, θ 3 are each calculated between two coordinates as explained before. The 4 th coordinate is added by pairing rotations with the first 3 coordinates, and so on.