Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14.

Slides:



Advertisements
Similar presentations
Chapter 13 Control Structures. Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display Control Structures Conditional.
Advertisements

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 301 Finite.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Chapter Four Trusses.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 One-Dimensional Unconstrained Optimization Chapter.
Optimization 吳育德.
5/20/ Multidimensional Gradient Methods in Optimization Major: All Engineering Majors Authors: Autar Kaw, Ali.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 61.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Ordinary Differential Equations Equations which are.
Optimization Methods One-Dimensional Unconstrained Optimization
Optimization Mechanics of the Simplex Method
Optimization Methods One-Dimensional Unconstrained Optimization
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 171 CURVE.
Advanced Topics in Optimization
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 11.
Optimization Methods One-Dimensional Unconstrained Optimization
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Part 31 Chapter.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Part 81 Partial.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Chapter 2 Image Slides.
Chapter 8 Traffic-Analysis Techniques. Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 8-1.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Numerical Differentiation and Integration Part 6 Calculus.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Part 31 Linear.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 111.
Nonlinear programming Unconstrained optimization techniques.
Chapter 7 Optimization. Content Introduction One dimensional unconstrained Multidimensional unconstrained Example.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 101.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 71.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 261 Stiffness.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 271 Boundary-Value.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Chapter 15 General Least Squares and Non- Linear.
Exam 1 Oct 3, closed book Place ITE 119, Time:12:30-1:45pm
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 2 - Chapter 7 Optimization.
Chapter 13 Transportation Demand Analysis. Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display
Introduction to Algorithms Second Edition by
Introduction to Algorithms Second Edition by
PowerPoint Presentations
Copyright © The McGraw-Hill Companies, Inc
Chapter 14.
Chapter 3 Image Slides Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
Introduction to Algorithms Second Edition by
Chapter 7 Optimization.
Chapter R A Review of Basic Concepts and Skills
Introduction to Algorithms Second Edition by
Introduction to Algorithms Second Edition by
Copyright © 2004 The McGraw-Hill Companies, Inc. All rights reserved.
Assignment Pages: 10 – 12 (Day 1) Questions over Assignment? # 1 – 4
Introduction to Algorithms Second Edition by
Chapter R.2 A Review of Basic Concepts and Skills
Chapter 12 Image Slides Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
Introduction to Algorithms Second Edition by
Introduction to Algorithms Second Edition by
Chapter 5 Foundations in Microbiology Fourth Edition
Copyright © 2004 The McGraw-Hill Companies, Inc. All rights reserved.
Introduction to Algorithms Second Edition by
Title Chapter 22 Image Slides
Copyright © The McGraw-Hill Companies, Inc
Copyright © 2004 The McGraw-Hill Companies, Inc. All rights reserved.
Introduction to Algorithms Second Edition by
CHAPTER 6 SKELETAL SYSTEM
Copyright © 2004 The McGraw-Hill Companies, Inc. All rights reserved.
Introduction to Algorithms Second Edition by
Part 2 Chapter 7 Optimization
Journey to the Cosmic Frontier
Journey to the Cosmic Frontier
Journey to the Cosmic Frontier
Copyright © 2004 The McGraw-Hill Companies, Inc. All rights reserved.
Copyright © 2004 The McGraw-Hill Companies, Inc. All rights reserved.
Chapter 3 Introduction to Physical Design of Transportation Facilities.
Introduction to Algorithms Second Edition by
Presentation transcript:

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Multidimensional Unconstrained Optimization Chapter 14 Techniques to find minimum and maximum of a function of several variables are described. These techniques are classified as: –That require derivative evaluation Gradient or descent (or ascent) methods –That do not require derivative evaluation Non-gradient or direct methods.

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Figure 14.1

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 DIRECT METHODS Random Search Based on evaluation of the function randomly at selected values of the independent variables. If a sufficient number of samples are conducted, the optimum will be eventually located. Example: maximum of a function f (x, y)=y-x-2x 2 -2xy-y 2 can be found using a random number generator.

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Figure 14.2

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Advantages/ Works even for discontinuous and nondifferentiable functions. Disadvantages/ As the number of independent variables grows, the task can become onerous. Not efficient, it does not account for the behavior of underlying function.

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Univariate and Pattern Searches More efficient than random search and still doesn’t require derivative evaluation. The basic strategy is: –Change one variable at a time while the other variables are held constant. –Thus problem is reduced to a sequence of one- dimensional searches that can be solved by variety of methods. –The search becomes less efficient as you approach the maximum.

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Figure 14.3

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 GRADIENT METHODS Gradients and Hessians The Gradient/ If f(x,y) is a two dimensional function, the gradient vector tells us –What direction is the steepest ascend? –How much we will gain by taking that step? Directional derivative of f(x,y) at point x=a and y=b

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 For n dimensions Figure 14.6

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 The Hessian/ For one dimensional functions both first and second derivatives valuable information for searching out optima. –First derivative provides (a) the steepest trajectory of the function and (b) tells us that we have reached the maximum. –Second derivative tells us that whether we are a maximum or minimum. For two dimensional functions whether a maximum or a minimum occurs involves not only the partial derivatives w.r.t. x and y but also the second partials w.r.t. x and y.

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Figure 14.7 Figure 14.8

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 Assuming that the partial derivatives are continuous at and near the point being evaluated The quantity [H] is equal to the determinant of a matrix made up of second derivatives

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 The Steepest Ascend Method Start at an initial point (x o,y o ), determine the direction of steepest ascend, that is, the gradient. Then search along the direction of the gradient, h o, until we find maximum. Process is then repeated. Figure 14.9

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 The problem has two parts –Determining the “best direction” and –Determining the “best value” along that search direction. Steepest ascent method uses the gradient approach as its choice for the “best” direction. To transform a function of x and y into a function of h along the gradient section: h is distance along the h axis

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14 If x o =1 and y o =2