1 OR II GSLM 52800. 2 Outline  some terminology  differences between LP and NLP  basic questions in NLP  gradient and Hessian  quadratic form  contour,

Slides:



Advertisements
Similar presentations
Chapter 11-Functions of Several Variables
Advertisements

Nonlinear Programming McCarl and Spreen Chapter 12.
Engineering Optimization
MS&E 211 Quadratic Programming Ashish Goel. A simple quadratic program Minimize (x 1 ) 2 Subject to: -x 1 + x 2 ≥ 3 -x 1 – x 2 ≥ -2.
Introducción a la Optimización de procesos químicos. Curso 2005/2006 BASIC CONCEPTS IN OPTIMIZATION: PART II: Continuous & Unconstrained Important concepts.
Optimization 吳育德.
Engineering Optimization
1 OR II GSLM Outline  classical optimization – unconstrained optimization  dimensions of optimization  feasible direction.
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
D Nagesh Kumar, IIScOptimization Methods: M2L2 1 Optimization using Calculus Convexity and Concavity of Functions of One and Two Variables.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Thursday, April 25 Nonlinear Programming Theory Separable programming Handouts: Lecture Notes.
Easy Optimization Problems, Relaxation, Local Processing for a small subset of variables.
Basic Feasible Solutions: Recap MS&E 211. WILL FOLLOW A CELEBRATED INTELLECTUAL TEACHING TRADITION.
Linear Discriminant Functions
The Most Important Concept in Optimization (minimization)  A point is said to be an optimal solution of a unconstrained minimization if there exists no.
MIT and James Orlin © Nonlinear Programming Theory.
OPTIMAL CONTROL SYSTEMS
Dual Problem of Linear Program subject to Primal LP Dual LP subject to ※ All duality theorems hold and work perfectly!
The Widrow-Hoff Algorithm (Primal Form) Repeat: Until convergence criterion satisfied return: Given a training set and learning rate Initial:  Minimize.
Constrained Optimization
1 Lecture 11 Resource Allocation Part1 (involving Continuous Variables- Linear Programming) 1.040/1.401/ESD.018 Project Management Samuel Labi and Fred.
Unconstrained Optimization Problem
Optimization using Calculus
Optimality Conditions for Nonlinear Optimization Ashish Goel Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
D Nagesh Kumar, IIScOptimization Methods: M2L3 1 Optimization using Calculus Optimization of Functions of Multiple Variables: Unconstrained Optimization.
Math for CSLecture 51 Function Optimization. Math for CSLecture 52 There are three main reasons why most problems in robotics, vision, and arguably every.
Tier I: Mathematical Methods of Optimization
1 OR II GSLM Outline  separable programming  quadratic programming.
Introduction to Optimization (Part 1)
Today Wrap up of probability Vectors, Matrices. Calculus
Computational Optimization
KKT Practice and Second Order Conditions from Nash and Sofer
1. Problem Formulation. General Structure Objective Function: The objective function is usually formulated on the basis of economic criterion, e.g. profit,
Linear Programming Piyush Kumar. Graphing 2-Dimensional LPs Example 1: x y Feasible Region x  0y  0 x + 2 y  2 y  4 x  3 Subject.
ENCI 303 Lecture PS-19 Optimization 2
Maxima and minima.
Response surfaces. We have a dependent variable y, independent variables x 1, x 2,...,x p The general form of the model y = f(x 1, x 2,...,x p ) +  Surface.
Nonlinear Programming Models
EASTERN MEDITERRANEAN UNIVERSITY Department of Industrial Engineering Non linear Optimization Spring Instructor: Prof.Dr.Sahand Daneshvar Submited.
Nonlinear Programming I Li Xiaolei. Introductory concepts A general nonlinear programming problem (NLP) can be expressed as follows: objective function.
Review of Matrix Operations Vector: a sequence of elements (the order is important) e.g., x = (2, 1) denotes a vector length = sqrt(2*2+1*1) orientation.
Signal & Weight Vector Spaces
Performance Surfaces.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Introduction to Optimization
1 OR II GSLM Outline  equality constraint  tangent plane  regular point  FONC  SONC  SOSC.
MA Day 19- February 1, 2013 Begin Differential Multivariable Calculus Section 11.1 Section 9.6.
Linear & Nonlinear Programming -- Basic Properties of Solutions and Algorithms.
A Kernel Approach for Learning From Almost Orthogonal Pattern * CIS 525 Class Presentation Professor: Slobodan Vucetic Presenter: Yilian Qin * B. Scholkopf.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
LINEAR PROGRAMMING Classic Types of Constrained Decision Problems 1. Product Mix 2. Ingredient Mix 3. Transportation 4. Assignment 5. Time Period.
Generalization Error of pac Model  Let be a set of training examples chosen i.i.d. according to  Treat the generalization error as a r.v. depending on.
OR II GSLM
D Nagesh Kumar, IISc Water Resources Systems Planning and Management: M2L2 Introduction to Optimization (ii) Constrained and Unconstrained Optimization.
Chapter 11 Optimization with Equality Constraints
Computational Optimization
Computational Optimization
Lecture 8 – Nonlinear Programming Models
Iterative Non-Linear Optimization Methods
1. Problem Formulation.
Chap 3. The simplex method
3-3 Optimization with Linear Programming
Derivative of scalar forms
L5 Optimal Design concepts pt A
EE 458 Introduction to Optimization
LINEARPROGRAMMING 4/26/2019 9:23 AM 4/26/2019 9:23 AM 1.
Performance Surfaces.
Part 3. Linear Programming
Outline Preface Fundamentals of Optimization
Presentation transcript:

1 OR II GSLM 52800

2 Outline  some terminology  differences between LP and NLP  basic questions in NLP  gradient and Hessian  quadratic form  contour, graph, and tangent plane

3 Feasible Points, Solution Set, and Neighborhood  feasible point: a point that satisfies all the constraints  solution set (feasible set, feasible region): the collection of all feasible points  neighborhood of x 0 = {x| |x  x 0 | <  } for some pre-specified  feasible region the neighborhood of a point for a given  A C B D  only the neighborhood of D is completely feasible for this 

4 Weak and Strong; Local and Global  local minima: x 1, any point in [s, t], x 3  strict (strong) local minima: x 1, x 3  weak local minima: any point in [s, t]  strict global minimum: x 1  weak local maxima: any point in [s, t] x f(x ) 12x3x3 x2x2 x1x1 t s

5 Differences Between Linear and Non-Linear Programming  linear programming  there exists an optimal extreme point (a corner point)  direction of improvement keeps on being so unless hitting a constraint  a local optimum point is also globally optimal direction of improvement optimal point

6 Differences Between Linear and Non-Linear Programming  none of these necessarily holds for a non- linear program x f(x ) 12x3x3 x2x2 x1x1 t s min x 2 + y 2, s.t. -2  x, y  2

7 Basic Questions in Non-Linear Programming  main question: given an initial location x 0, how to get to a local minimum, or, better, a global minimum  (a) the direction of improvement?  (b) the necessary conditions of an optimal point?  (c) the sufficient conditions of an optimal point?  (d) any conditions to simplify the processes in (a), (b), and (c)?  (e) any algorithmic procedures to solve a NLP problem?

8 Basic Questions in Non-Linear Programming  calculus required for (a) to (e)  direction of improvement of f = gradient of f  shaped by constraints  convexity for (d), and also (b) and (c)  identification of convexity: definiteness of matrices, especially for Hessians

9 Gradient and Hessian  gradient of f:  f(x) =  in short  Hessian = f and g j usually assumed to be twice differentiable functions  Hessian is a symmetric matrix

10 Gradient and Hessian  e j : (0, …, 0, 1, 0, …, 0) T, where “1” at the jth position   for small , f(x+  e j )  f(x) +   in general,  x = (  x 1, …,  x n ) T from x, f(x+  x)  f(x) +

11 Example  (a). f(x) = x 2 ; f(3.5+  )  ? for small   (b). f(x, y) = x 2 + y 2, f((1, 1) + (  x,  y))  ? for small  x,  y  gradient  f : direction of steepest accent of the objective fucntion

12 Example  find the Hessian of  (a). f(x, y) = x 2 + 7y 2  (b). f(x, y) = x 2 + 5xy + 7y 2  (c). f(x, y) = x 3 + 7y 2

13 Quadratic Form  general form: x T Qx/2 + c T x + a, where x is an n-dimensional vector; Q an n  n square matrix; c and a are matrices of appropriate dimensions  how to derive the gradient and Hessian?  gradient  f(x) = Qx+c  Hessian H = Q

14 Quadratic Form  relate the two forms x T Qx/2 + c T x + a and f(x, y) =  1 x 2 +  2 xy+  3 y 2 +  4 x+  5 y+  6  Example 1.6.3

15 Example  Find the first two derivatives of the following f(x)  f(x) = x 2 for x  [-2, 2]  f(x) = -x 2 for x  [-2, 2]

16 Contour and Graph (i.e., Surface) of Function f  Example 1.7.1: f(x 1, x 2 ) =

17 Contour and Graph (i.e., Surface) of Function f  an n-dimensional function  a contour of f: a diagram f(x) = c in the n- dimensional space for a given value c  the graph (surface function) of f: the diagram z = f(x) in the (n+1)st dimensional space as x and z vary

18 Contour and Graph (i.e., Surface) of Function f  how do the contours of the one-dimensional function f(x) = x 2 look like?

19 An Important Property Between the Gradient and the Tangent Plane at a Contour  the gradient of f at point x 0 is orthogonal to the tangent of the contour f(x) = c at x 0  many optimization results are related to the above property

20 Gradient of f at x 0 Being Orthogonal to the Tangent of the Contour f(x) = c at x 0  Example 1.7.3: f(x 1, x 2 ) = x 1 +2x 2  gradient at (4, 2)?  tangent of contour at (4, 2)?

21 Gradient of f at x 0 Being Orthogonal to the Tangent of the Contour f(x) = c at x 0  Example 1.7.2: f(x 1, x 2 ) =  point (x 10, x 20 ) on a contour f(x 1, x 2 ) = c

22 Tangent at a Contour and the Corresponding Tangent Plane at a Surface  the above two are related  for contour of f(x, y) = x 2 +y 2, the tangent at (x 0, y 0 )  (x-x 0, y- y 0 ) T (2x 0, 2y 0 ) = 0 two orthogonal vectors u and v: u T v = 0

23 Tangent at a Contour and the Corresponding Tangent Plane at a Surface  the tangent place at (x 0, y 0 ) for the surface of f(x, y) = x 2 +y 2  the surface: z = x 2 +y 2  defining a contour at a higher dimension: F(x, y, z) = x 2 +y 2  z  tangent plane at (x 0, y 0, ) of the surface: what happens when z =