CS5321 Numerical Optimization

Slides:



Advertisements
Similar presentations
Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
Advertisements

Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Solving Linear Systems (Numerical Recipes, Chap 2)
Symmetric Matrices and Quadratic Forms
3D Geometry for Computer Graphics
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Chapter 2 Matrices Definition of a matrix.
3D Geometry for Computer Graphics
Ordinary least squares regression (OLS)
5 5.1 © 2012 Pearson Education, Inc. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
化工應用數學 授課教師: 郭修伯 Lecture 9 Matrices
A matrix having a single row is called a row matrix. e.g.,
5.1 Orthogonality.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
CHAPTER SIX Eigenvalues
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
Eigenvalue Problems Solving linear systems Ax = b is one part of numerical linear algebra, and involves manipulating the rows of a matrix. The second main.
8.1 Vector spaces A set of vector is said to form a linear vector space V Chapter 8 Matrices and vector spaces.
Square n-by-n Matrix.
Matrices CHAPTER 8.1 ~ 8.8. Ch _2 Contents  8.1 Matrix Algebra 8.1 Matrix Algebra  8.2 Systems of Linear Algebra Equations 8.2 Systems of Linear.
Matrices. Definitions  A matrix is an m x n array of scalars, arranged conceptually as m rows and n columns.  m is referred to as the row dimension.
Linear algebra: matrix Eigen-value Problems
Computing Eigen Information for Small Matrices The eigen equation can be rearranged as follows: Ax = x  Ax = I n x  Ax - I n x = 0  (A - I n )x = 0.
Linear algebra: matrix Eigen-value Problems Eng. Hassan S. Migdadi Part 1.
Chapter 5 MATRIX ALGEBRA: DETEMINANT, REVERSE, EIGENVALUES.
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 07 Chapter 7: Eigenvalues, Eigenvectors.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Review of Linear Algebra Optimization 1/16/08 Recitation Joseph Bradley.
Review of Matrix Operations Vector: a sequence of elements (the order is important) e.g., x = (2, 1) denotes a vector length = sqrt(2*2+1*1) orientation.
5.1 Eigenvectors and Eigenvalues 5. Eigenvalues and Eigenvectors.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
5 5.1 © 2016 Pearson Education, Ltd. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
2 - 1 Chapter 2A Matrices 2A.1 Definition, and Operations of Matrices: 1 Sums and Scalar Products; 2 Matrix Multiplication 2A.2 Properties of Matrix Operations;
Boot Camp in Linear Algebra TIM 209 Prof. Ram Akella.
Matrices, Vectors, Determinants.
Reduced echelon form Matrix equations Null space Range Determinant Invertibility Similar matrices Eigenvalues Eigenvectors Diagonabilty Power.
CS246 Linear Algebra Review. A Brief Review of Linear Algebra Vector and a list of numbers Addition Scalar multiplication Dot product Dot product as a.
Matrices and Vector Concepts
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Review of Linear Algebra
CS479/679 Pattern Recognition Dr. George Bebis
Review of Matrix Operations
Elementary Linear Algebra Anton & Rorres, 9th Edition
Matrices and vector spaces
Systems of First Order Linear Equations
Euclidean Inner Product on Rn
Singular Value Decomposition
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Some useful linear algebra
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS485/685 Computer Vision Dr. George Bebis
~ Least Squares example
Symmetric Matrices and Quadratic Forms
Linear Algebra Lecture 32.
Maths for Signals and Systems Linear Algebra in Engineering Lecture 18, Friday 18th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
~ Least Squares example
Elementary Linear Algebra Anton & Rorres, 9th Edition
Eigenvalues and Eigenvectors
Linear Algebra: Matrix Eigenvalue Problems – Part 2
Lecture 20 SVD and Its Applications
Linear Algebra Lecture 28.
Symmetric Matrices and Quadratic Forms
CSE 203B: Convex Optimization Week 2 Discuss Session
CS5321 Numerical Optimization
Presentation transcript:

CS5321 Numerical Optimization 01 Introduction 7/30/2019

Class information Class webpage: http://www.cs.nthu.edu.tw/∼cherung/cs5321 Text and reference books: Numerical optimization, Jorge Nocedal and Stephen J. Wright (http://www.mcs.anl.gov/otc/Guide) Linear and Nonlinear Programming, Stephen G. Nash and Ariela Sofer (1996, 2005) Numerical Methods for Unconstrained Optimization and Nonlinear Equations, J. Dennis and R. Schnabel TA: 王治權 pponywong@gmail.com 7/30/2019

Grading Class notes (50%) Project (40%) Class attendance (10%) Using latex to write one class note. Peer review system You must review others’ notes and give comments The grade is given on both works (30%-20%) Project (40%) Applications, software survey, implementations Proposal (10%), presentation(10%), and report (20%) Class attendance (10%) 7/30/2019

Outline of class Introduction Unconstrained optimization Background Fundamental of unconstrained optimization Line search methods Trust region methods Conjugate gradient methods Quasi-Newton methods Inexact Newton methods 7/30/2019

Outline of class-continue Linear programming The simplex method The interior point method Constrained optimization Optimality conditions Quadratic programming Penalty and augmented Largrangian methods Active set methods Interior point methods 7/30/2019

Optimization Tree 7/30/2019

Classification Minimization or maximization Continuous vs. discrete optimization Constrained and unconstrained optimization Linear and nonlinear programming Convex and non-convex problems Global and local solution 7/30/2019

Affine, convex, and cones where xiRn and iR . x is a linear combination of {xi} x is an affine combination of {xi} if x is a conical combination of {xi} if x is a convex combination of {xi} if p X i = 1 ¸ ¸ i p X i = 1 ¸ ; 7/30/2019

linear combination conical combination affine combination convex combination 7/30/2019

Convex function A function f is convex if for [0,1] A function f is concave if −f is convex 7/30/2019

Some calculus Rate of convergence Lipschitz continuity Single valued function Derivatives and Hessian Mean value theorem and Taylor’s theorem Vector valued function Derivatives and Jacobian 7/30/2019

Rate of convergence Let {xk|xkRn} be a sequence converge to x* The convergence is Q-linear if for a constant r(0,1) The convergence is Q-superlinear if The convergence is p Q-order if for a constant M 7/30/2019

|| f (x) − f (y)||  L ||x − y|| for all x,yN Lipschitz continuity A function f is said to be Lipschitz continuous on some set N if there is a constant L>0 such that || f (x) − f (y)||  L ||x − y|| for all x,yN If function f and g are Lipschitz continuous on N, f +g and fg are Lipschitz continuous on N. 7/30/2019

Derivatives For a single valued function f (x1,…,xn):RnR Partial derivative of xi: Gradient of f is Directional derivatives: for ||p||=1 7/30/2019

Hessian In some sense of the second derivative of f If f is twice continuously differentiable, Hessian is symmetric. 7/30/2019

Taylor’s theorem Mean value theorem: for (0,1) Taylor’s theorem: for (0,1) f ( x + p ) = r T 1 2 ® 7/30/2019

Vector valued function For a vector valued function r:RnRm The Jacobian of r at x is 7/30/2019

Some linear algebra Vectors and matrix Eigenvalue and eigenvector Singular value decomposition LU decomposition and Cholesky decomposition Subspaces and QR decomposition Sherman-Morrison-Woodbury formula 7/30/2019

Vector A column vector xRn is denoted as The transpose of x is The inner product of x,yRn is Vector norm 1-norm 2-norm -norm 7/30/2019

Matrix A matrix ARmn is The transpose of A is Matrix A is symmetric if AT=A Matrix norm: ||A||p= max||Ax||p for ||x||p=1, p=1,2, 7/30/2019

Eigenvalue and eigenvector A scalar  is an eigenvalue of an nn matrix A if there is a nonzero vector x such that Ax= x. Vector x is called an eigenvector. Matrix A is symmetric positive definite (SPD) if AT=A and all its eigenvalues are positive. If A has n linearly independent eigenvectors, A can have the eigen-decomposition: A=XX −1.  is diagonal with eigenvalues as its diagonal elements Column vectors of X are corresponding eigenvectors 7/30/2019

Spectral decomposition If A is real and symmetric, all its eigenvalues are real, and there are n orthogonal eigenvectors. The spectral decomposition of a symmetric matrix A is A=QQT.  is diagonal with eigenvalues as its diagonal elements Q is orthogonal, i.e. QTQ = QQT = I. Column vectors of Q are corresponding eigenvectors. 7/30/2019

Singular value The singular values of an mn A are the square roots of the eigenvalues of ATA. Any matrix A can have the singular value decomposition (SVD): A=U VT.  is diagonal with singular values as its elements. U and V are orthogonal matrices. The column vectors of U are called left singular vectors of A; the column vectors of V is called the right singular vector of A. 7/30/2019

LU decomposition The LU decomposition with pivoting of matrix A is PA=LU P is a permutation matrix L is lower triangular; U is upper triangular. The linear system Ax=b can be solved by Perform LU decompose PA=LU Solve Ly=Pb Solve Ux=y 7/30/2019

Cholesky decomposition For a SPD matrix A, there exists the Cholesky decomposition PTAP = LLT P is a permutation matrix L is a lower triangular matrix If A is not SPD, the LBL decomposition can be used: PTAP = LBLT B is a block diagonal matrix with blocks of dimension 1 or 2. 7/30/2019

Subspaces, QR decomposition The null space of an mn matrix A is Null(A)={w|Aw=0,w0} The range of A is Range(A)={w|w=Av,v}. Fundamental of linear algebra: Null(A)  Range(AT) = Rn Matrix A has the QR decomposition AP = QR P is permutation matrix; Q is an orthogonal matrix; R is an upper triangular matrix. 7/30/2019

Singularity and ill-conditioned An nn matrix A is singular (noninvertible) iff A has 0 eigenvalues A has 0 singular values The null space of A is not empty The determinant of A is zero The condition number of A is A is ill-conditioned if it has a large condition number. · ( A ) = k ¡ 1 7/30/2019

Sherman-Morrison-Woodbury formula For a nonsinular matrix ARnn, if a rank-one update of  =A+uvT is also nonsingular, For matrix U,VRnp, 1pn, if  =A+UVT is nonsingular, 7/30/2019