Constrained Optimization – Part 1

Slides:



Advertisements
Similar presentations
Fin500J: Mathematical Foundations in Finance
Advertisements

Linear Programming. Introduction: Linear Programming deals with the optimization (max. or min.) of a function of variables, known as ‘objective function’,
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
Optimality conditions for constrained local optima, Lagrange multipliers and their use for sensitivity of optimal solutions Today’s lecture is on optimality.
Chapter 3 Linear Programming: Sensitivity Analysis and Interpretation of Solution MT 235.
Finance 30210: Managerial Economics Optimization.
Optimization using Calculus
Linear Programming. Linear programming A technique that allows decision makers to solve maximization and minimization problems where there are certain.
1 Linear Programming Using the software that comes with the book.
Nonlinear Optimization Review of Derivatives Models with One Decision Variable Unconstrained Models with More Than One Decision Variable Models with Equality.
THE ABSOLUTE VALUE FUNCTION. Properties of The Absolute Value Function Vertex (2, 0) f (x)=|x -2| +0 vertex (x,y) = (-(-2), 0) Maximum or Minimum? a =
Introduction to Optimization (Part 1)
1 1 Slide LINEAR PROGRAMMING: THE GRAPHICAL METHOD n Linear Programming Problem n Properties of LPs n LP Solutions n Graphical Solution n Introduction.
1 Least Cost System Operation: Economic Dispatch 2 Smith College, EGR 325 March 10, 2006.
Optimization Techniques Methods for maximizing or minimizing an objective function Examples –Consumers maximize utility by purchasing an optimal combination.
Managerial Economics Managerial Economics = economic theory + mathematical eco + statistical analysis.
Chapter 19 Linear Programming McGraw-Hill/Irwin
Linear Programming Chapter 13 Supplement.
Optimization of Process Flowsheets S,S&L Chapter 24 T&S Chapter 12 Terry A. Ring CHEN 5253.
3.4 Linear Programming p Optimization - Finding the minimum or maximum value of some quantity. Linear programming is a form of optimization where.
Linear Programming: Sensitivity Analysis and Interpretation of Solution Pertemuan 5 Matakuliah: K0442-Metode Kuantitatif Tahun: 2009.
Slide 1  2002 South-Western Publishing Web Chapter A Optimization Techniques Overview Unconstrained & Constrained Optimization Calculus of one variable.
Managerial Economics Managerial Economics = economic theory + mathematical eco + statistical analysis.
D Nagesh Kumar, IIScOptimization Methods: M2L4 1 Optimization using Calculus Optimization of Functions of Multiple Variables subject to Equality Constraints.
Optimization unconstrained and constrained Calculus part II.
Calculus: Hughs-Hallett Chap 5 Joel Baumeyer, FSC Christian Brothers University Using the Derivative -- Optimization.
Optimization Problems
1 Optimization Techniques Constrained Optimization by Linear Programming updated NTU SY-521-N SMU EMIS 5300/7300 Systems Analysis Methods Dr.
Calculus-Based Optimization AGEC 317 Economic Analysis for Agribusiness and Management.
Section 15.3 Constrained Optimization: Lagrange Multipliers.
1 Unconstrained and Constrained Optimization. 2 Agenda General Ideas of Optimization Interpreting the First Derivative Interpreting the Second Derivative.
Operations Research By: Saeed Yaghoubi 1 Graphical Analysis 2.
D Nagesh Kumar, IISc Water Resources Systems Planning and Management: M2L2 Introduction to Optimization (ii) Constrained and Unconstrained Optimization.
1 2 Linear Programming Chapter 3 3 Chapter Objectives –Requirements for a linear programming model. –Graphical representation of linear models. –Linear.
Solver Feature Excel’s KY San Jose State University Engineering 10.
An Introduction to Linear Programming
Linear Programming for Solving the DSS Problems
Chapter 14 Partial Derivatives
Another sufficient condition of local minima/maxima
Solver Feature Excel’s KY San Jose State University Engineering 10.
Do Now from 1.2a Find the domain of the function algebraically and support your answer graphically. Find the range of the function.
deterministic operations research
An Introduction to Linear Programming Pertemuan 4
Chapter 11 Optimization with Equality Constraints
Calculus-Based Solutions Procedures MT 235.
Functions of Several Variables
Lecture 8 – Nonlinear Programming Models
Constrained Optimization
Unconstrained and Constrained Optimization
3-3 Optimization with Linear Programming
Duality Theory and Sensitivity Analysis
AP Calculus BC September 30, 2016.
Optimization Problems
Functions Introduction.
Constrained Optimization
Constrained Optimization – Part 1
7.5 – Constrained Optimization: The Method of Lagrange Multipliers
Chapter 5. The Duality Theorem
Outline Unconstrained Optimization Functions of One Variable
More Properties of Functions
EEE 244-8: Optimization.
Chapter 7 Functions of Several Variables
Graphical solution A Graphical Solution Procedure (LPs with 2 decision variables can be solved/viewed this way.) 1. Plot each constraint as an equation.
Calculus I (MAT 145) Dr. Day Wednesday March 27, 2019
Constrained Optimization
Calculus-Based Optimization AGEC 317
Multivariable optimization with no constraints
Maximum and Minimum Values
Analyzing Multivariable Change: Optimization
Presentation transcript:

Constrained Optimization – Part 1 Objective of Presentation: To introduce Lagrangean as a basic conceptual method used to optimize design in real situations Essential Reality: In practical situations, the designers are constrained or limited by physical realities design standards laws and regulations, etc.

Outline Unconstrained Optimization (Review) Constrained Optimization – Lagrangeans Approach Lagrangeans as Equality constraints Interpretation of Lagrangeans as “Shadow prices”

Unconstrained Optimization: Definitions Optimization => Maximum of desired quantity, or => Minimum of undesired quantity Objective Function = Formula to be optimized = Z(X) Decision Variables = Variables about which we can make decisions = X = (X1….Xn)

Unconstrained Optimization: Graph B C D E F(X) X B and D are maxima A, C and E are minima

Unconstrained Optimization: Conditions By calculus: if F(X) continuous, analytic Primary conditions for maxima and minima: F(X) / Xi = 0 i ( symbol means: “for all i”) Secondary conditions: 2F(X) / Xi2 < 0 = > Max (B,D) 2F(X) / Xi2 > 0 = > Min (A,C,E) These define whether point of no change in Z is a maximum or a minimum

Unconstrained Optimization: Example Example: Housing insulation Total Cost = Fuel cost + Insulation cost x = Thickness of insulation F(x) = K1 / x + K2x Primary condition: F(x) / x = 0 = -K1 / x2 + K2 => x* = {K1 / K2} 1/2 (starred quantities are optimal)

Unconstrained Optimization: Graph of Solution to Example If: K1 = 500 ; K2 = 24 Then: X* = 4.56

Constrained Optimization: General “Constrained Optimization” involves the optimization of a process subject to constraints Constraints have two basic types Equality Constraints -- some factors have to equal constraints Inequality Constraints -- some factors have to be less less or greater than the constraints (these are “upper” and “lower” bounds)

Constrained Optimization: General Approach To solve situations of increasing complexity, (for example, those with equality, inequality constraints) … Transform more difficult situation into one we know how to deal with Note: this process introduces new variables! Thus, transform “constrained” optimization to “unconstrained” optimization

Equality Constraints: Example Example: Best use of budget Maximize: Output = Z(X) = a0x1a1x2a2 Subject to (s.t.): Total costs = Budget = p1x1 + p2x2 Z(x) Z* Budget X Note: Z(X) / X  0 at optimum

Lagrangean Method: Approach Transforms equality constraints into unconstrained problem Start with: Opt: F(x) s.t.: gj(x) = bj => gj(x) - bj = 0 Get to: L = F(x) - j j[gj(x) - bj] j = Lagrangean multipliers (lambdas sub j) -- are unknown quantities for which we must solve Note: [gj(x) - bj] = 0 by definition, thus optimum for F(x) = optimum for L

Lagrangean: Optimality Conditions Since the new formulation is a single equation, we can use formulas for unconstrained optimization. We set partial derivatives equal to zero for all unknowns, the X and the  Thus, to optimize L: L / xi = 0 I L / j = 0 J

Lagrangean: Example Formulation Problem: Opt: F(x) = 6x1x2 s.t.: g(x) = 3x1 + 4x2 = 18 Lagrangean: L = 6x1x2 - (3x1 + 4x2 - 18) Optimality Conditions: L / x1 = 6x2 - 3 = 0 L / x2 = 6x1 - 4 = 0 L / j = 3x1 + 4x2 -18 = 0

Lagrangean: Graph for Example

Lagrangean: Example Solution Solving as unconstrained problem: L / x1 = 6x2 - 3 = 0 L / x2 = 6x1 - 4 = 0 L / i = 3x1 + 4x2 -18 = 0 so that:  = 2x2 = 1.5x1 (first 2 equations) => x2 = 0.75x1 => 3x1 + 3x1 - 18 = 0 (3rd equation) x1* = 18/6 = 3 x2* = 18/8 = 2.25 * = 4.5 F(x)* = 40.5

Lagrangean: Graph for Solution 

Shadow Price Shadow Price = Rate of change of objective function per unit change of constraint = F(x) / bj = j It is extremely important for system design It defines value of changing constraints, and indicates if worthwhile to change them Should we buy more resources? Should we change environmental constraints?

Lagrangean Multiplier is a Shadow Price The Lagrangean multiplier is interpreted as the shadow price on constraint SPj = F(x)*/ bj = L*/ bj =  {F(x) - j j[gj(x) - bj] } / bj = j Naturally, this is an instantaneous rate

Lagrangean = Shadow Prices Example Let’s see how this works in example, by changing constraint by 0.1 units: Opt: F(x) = 6x1x2 s.t.: g(x) = 3x1 + 4x2 = 18.1 The optimum values of the variables are x1* = (18.1)/6 x2* = (18.1)/8 * = 4.5 Thus F(x)* = 6(18.1/6)(18.1/8) = 40.95 F(x) = 40.95 - 40.5 = 0.45 = * (0.1)

Generalization In general constraints are “inequalities”: Upper bounds : gj(x) < bj Lower bounds: gj(x) > bj At optimum, some constraints will limit solution (they are “binding”) others not Example: airline bags: weight < 40kg ; sum of dimensions < 2.5m. Your bag might be limited by weight, not by size. Shadow prices > 0 for all “binding” constraints = 0 for all others

Design Implications Expanding range of design variables (x), increases freedom to improve design, thus adds value This is called “relaxing” the constraints Increasing upper bounds Decreasing lower bounds As any constraint relaxed, it may no longer be “binding” , and others can become so SHADOW PRICES DEPEND ON OTHER CONSTRAINTS, “PROBLEM DEPENDENT”

Take-aways Relaxing design constraints adds value (in terms of better performance, F(x) ) This value is the “shadow price” of that constraint Knowing this can be very important for designers, shows way to improve quickly NOTE: Value to design has no direct connection to cost of constraint, not a “price in ordinary terms