Steepest Decent and Conjugate Gradients (CG). Solving of the linear equation system.

Slides:



Advertisements
Similar presentations
Instabilities of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small.
Advertisements

Optimization.
Optimization : The min and max of a function
Optimization of thermal processes
Optimization 吳育德.
Solving Linear Systems (Numerical Recipes, Chap 2)
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
Linear Discriminant Functions
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
1cs542g-term Notes  In assignment 1, problem 2: smoothness = number of times differentiable.
Jonathan Richard Shewchuk Reading Group Presention By David Cline
1cs542g-term Notes  Assignment 1 due tonight ( me by tomorrow morning)
Gradient Methods April Preview Background Steepest Descent Conjugate Gradient.
Numerical Optimization
Function Optimization Newton’s Method. Conjugate Gradients
1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me .
Tutorial 12 Unconstrained optimization Conjugate gradients.
Optimization Methods One-Dimensional Unconstrained Optimization
Gradient Methods May Preview Background Steepest Descent Conjugate Gradient.
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Optimization Methods One-Dimensional Unconstrained Optimization
Gradient Methods Yaron Lipman May Preview Background Steepest Descent Conjugate Gradient.
Function Optimization. Newton’s Method Conjugate Gradients Method
CSE 245: Computer Aided Circuit Simulation and Verification
Optimization Methods One-Dimensional Unconstrained Optimization

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
Introduction to Numerical Analysis I MATH/CMPSC 455 Conjugate Gradient Methods.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Numerical Analysis 1 EE, NCKU Tien-Hao Chang (Darby Chang)
Computational Optimization
UNCONSTRAINED MULTIVARIABLE
Point set alignment Closed-form solution of absolute orientation using unit quaternions Berthold K. P. Horn Department of Electrical Engineering, University.
Systems of Linear Equations Iterative Methods
ENCI 303 Lecture PS-19 Optimization 2
Nonlinear least squares Given m data points (t i, y i ) i=1,2,…m, we wish to find a vector x of n parameters that gives a best fit in the least squares.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
CSE 245: Computer Aided Circuit Simulation and Verification Matrix Computations: Iterative Methods I Chung-Kuan Cheng.
559 Fish 559; Lecture 5 Non-linear Minimization. 559 Introduction Non-linear minimization (or optimization) is the numerical technique that is used by.
Case Study in Computational Science & Engineering - Lecture 5 1 Iterative Solution of Linear Systems Jacobi Method while not converged do { }
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Chapter 10 Minimization or Maximization of Functions.
Exam 1 Oct 3, closed book Place ITE 119, Time:12:30-1:45pm
Gradient Methods In Optimization
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Krylov-Subspace Methods - I Lecture 6 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Linear Systems Numerical Methods. 2 Jacobi Iterative Method Choose an initial guess (i.e. all zeros) and Iterate until the equality is satisfied. No guarantee.
Numerical Analysis – Data Fitting Hanyang University Jong-Il Park.
MA5233 Lecture 6 Krylov Subspaces and Conjugate Gradients Wayne M. Lawton Department of Mathematics National University of Singapore 2 Science Drive 2.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Krylov-Subspace Methods - II Lecture 7 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Krylov-Subspace Methods - I
Non-linear Minimization
CSE 245: Computer Aided Circuit Simulation and Verification
Non-linear Least-Squares
CSE 245: Computer Aided Circuit Simulation and Verification
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
Conjugate Gradient Method
CS5321 Numerical Optimization
Introduction to Scientific Computing II
Introduction to Scientific Computing II
Introduction to Scientific Computing II
~ Least Squares example
~ Least Squares example
Introduction to Scientific Computing II
Performance Optimization
Outline Preface Fundamentals of Optimization
Section 3: Second Order Methods
Conjugate Direction Methods
Presentation transcript:

Steepest Decent and Conjugate Gradients (CG)

Solving of the linear equation system

Steepest Decent and Conjugate Gradients (CG) Solving of the linear equation system Problem: dimension n too big, or not enough time for gauss elimination Iterative methods are used to get an approximate solution.

Steepest Decent and Conjugate Gradients (CG) Solving of the linear equation system Problem: dimension n too big, or not enough time for gauss elimination Iterative methods are used to get an approximate solution. Definition Iterative method: given starting point, do steps hopefully converge to the right solution

starting issues

Solving is equivalent to minimizing

starting issues Solving is equivalent to minimizing A has to be symmetric positive definite:

starting issues

starting issues If A is also positive definite the solution of is the minimum

starting issues If A is also positive definite the solution of is the minimum

starting issues error: The norm of the error shows how far we are away from the exact solution, but can’t be computed without knowing of the exact solution.

starting issues error: The norm of the error shows how far we are away from the exact solution, but can’t be computed without knowing of the exact solution. residual: can be calculated

Steepest Decent

We are at the point. How do we reach ?

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( )

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go?

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go? Choose so that is minimized:

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go? Choose so that is minimized:

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go? Choose so that is minimized:

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go? Choose so that is minimized:

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go? Choose so that is minimized:

Steepest Decent We are at the point. How do we reach ? Idea: go into the direction in which decreases most quickly ( ) how far should we go? Choose so that is minimized:

Steepest Decent one step of steepest decent can be calculated as follows:

Steepest Decent one step of steepest decent can be calculated as follows: stopping criterion: or with an given small It would be better to use the error instead of the residual, but you can’t calculate the error.

Steepest Decent Method of steepest decent:

Steepest Decent As you can see the starting point is important!

Steepest Decent As you can see the starting point is important! When you know anything about the solution use it to guess a good starting point. Otherwise you can choose a starting point you want e.g..

Steepest Decent - Convergence

Definition energy norm:

Steepest Decent - Convergence Definition energy norm: Definition condition: ( is the largest and the smallest eigenvalue of A)

Steepest Decent - Convergence Definition energy norm: Definition condition: ( is the largest and the smallest eigenvalue of A) convergence gets worse when the condition gets larger

Conjugate Gradients

is there a better direction?

Conjugate Gradients is there a better direction? Idea: orthogonal search directions

Conjugate Gradients is there a better direction? Idea: orthogonal search directions

Conjugate Gradients is there a better direction? Idea: orthogonal search directions only walk once in each direction and minimize

Conjugate Gradients is there a better direction? Idea: orthogonal search directions only walk once in each direction and minimize maximal n steps are needed to reach the exact solution

Conjugate Gradients is there a better direction? Idea: orthogonal search directions only walk once in each direction and minimize maximal n steps are needed to reach the exact solution has to be orthogonal to

Conjugate Gradients example with the coordinate axes as orthogonal search directions:

Conjugate Gradients example with the coordinate axes as orthogonal search directions: Problem: can’t be computed because (you don’t know !)

Conjugate Gradients new idea: A-orthogonal

Conjugate Gradients new idea: A-orthogonal Definition A-orthogonal: A-orthogonal (reminder: orthogonal: )

Conjugate Gradients new idea: A-orthogonal Definition A-orthogonal: A-orthogonal (reminder: orthogonal: ) now has to be A-orthogonal to

Conjugate Gradients new idea: A-orthogonal Definition A-orthogonal: A-orthogonal (reminder: orthogonal: ) now has to be A-orthogonal to

Conjugate Gradients new idea: A-orthogonal Definition A-orthogonal: A-orthogonal (reminder: orthogonal: ) now has to be A-orthogonal to can be computed!

Conjugate Gradients A set of A-orthogonal directions can be found with n linearly independent vectors and conjugate Gram- Schmidt (same idea as Gram-Schmidt).

Conjugate Gradients Gram-Schmidt: linearly independent vectors

Conjugate Gradients Gram-Schmidt: linearly independent vectors

Conjugate Gradients Gram-Schmidt: linearly independent vectors conjugate Gram-Schmidt:

Conjugate Gradients A set of A-orthogonal directions can be found with n linearly independent vectors and conjugate Gram- Schmidt (same idea as Gram-Schmidt). CG works by setting (makes conjugate Gram- Schmidt easy)

Conjugate Gradients A set of A-orthogonal directions can be found with n linearly independent vectors and conjugate Gram- Schmidt (same idea as Gram-Schmidt). CG works by setting (makes conjugate Gram- Schmidt easy) with

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Conjugate Gradients

Method of Conjugate Gradients:

Conjugate Gradients - Convergence

Conjugate Gradients - Convergence for steepest decent for CG Convergence of CG is much better!