Download presentation
Presentation is loading. Please wait.
Published byDonald Stanley Modified over 9 years ago
1
Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal January 31, 2007 Inexact Methods for PDE-Constrained Optimization University of North Carolina at Chapel Hill
2
Nonlinear Optimization “One” problem
3
Circuit Tuning Building blocks: Transistors (switches) and Gates (logic units) Improve aspects of the circuit – speed, area, power – by choosing transistor widths AT1 AT3 AT2 d1 d2 w1w2 (A. Wächter, C. Visweswariah, and A. R. Conn, 2005)
4
Circuit Tuning Building blocks: Transistors (switches) and Gates (logic units) Improve aspects of the circuit – speed, area, power – by choosing transistor widths Formulate an optimization problem AT1 AT3 AT2 d1 d2 w1w2 (A. Wächter, C. Visweswariah, and A. R. Conn, 2005)
5
Strategic Bidding Electricity production companies “bid” on how much they will charge for one unit of electricity Independent operator collects bids and sets production schedule and “spot price” to minimize cost to consumers (Pereira, Granville, Dix, and Barroso, 2004)
6
Strategic Bidding Electricity production companies “bid” on how much they will charge for one unit of electricity Independent operator collects bids and sets production schedule and “spot price” to minimize cost to consumers Bilevel problem Equivalent to MPCC Hard geometry! (Pereira, Granville, Dix, and Barroso, 2004)
7
Challenges for NLP algorithms Very large problems Numerical noise Availability of derivatives Degeneracies Difficult geometries Expensive function evaluations Real-time solutions needed Integer variables Negative curvature
8
Outline Problem Formulation Equality constrained optimization Sequential Quadratic Programming Inexact Framework Unconstrained optimization and nonlinear equations Stopping conditions for linear solver Global Behavior Merit function and sufficient decrease Satisfying first order conditions Numerical Results Model inverse problem Accuracy tradeoffs Final Remarks Future work Negative curvature
9
Outline Problem Formulation Equality constrained optimization Sequential Quadratic Programming Inexact Framework Unconstrained optimization and nonlinear equations Stopping conditions for linear solver Global Behavior Merit function and sufficient decrease Satisfying first order conditions Numerical Results Model inverse problem Accuracy tradeoffs Final Remarks Future work Negative curvature
10
Equality constrained optimization e.g., minimize the difference between observed and expected behavior, subject to atmospheric flow equations (Navier-Stokes) Goal: solve the problem
11
Equality constrained optimization Define: the Lagrangian Define: the derivatives Goal: solve KKT conditions
12
Sequential Quadratic Programming (SQP) Algorithm: Newton’s methodAlgorithm: the SQP subproblem Two “equivalent” step computation techniques
13
Sequential Quadratic Programming (SQP) Algorithm: Newton’s methodAlgorithm: the SQP subproblem Two “equivalent” step computation techniques KKT matrix Cannot be formed Cannot be factored
14
Sequential Quadratic Programming (SQP) Algorithm: Newton’s methodAlgorithm: the SQP subproblem Two “equivalent” step computation techniques KKT matrix Cannot be formed Cannot be factored Linear system solve Iterative method Inexactness
15
Outline Problem Formulation Equality constrained optimization Sequential Quadratic Programming Inexact Framework Unconstrained optimization and nonlinear equations Stopping conditions for linear solver Global Behavior Merit function and sufficient decrease Satisfying first order conditions Numerical Results Model inverse problem Accuracy tradeoffs Final Remarks Future work Negative curvature
16
Unconstrained optimization Goal: minimize a nonlinear objective Algorithm: Newton’s method (CG)
17
Unconstrained optimization Goal: minimize a nonlinear objective Algorithm: Newton’s method (CG) Note: choosing any intermediate step ensures global convergence to a local solution of NLP (Steihaug, 1983)
18
Nonlinear equations Goal: solve a nonlinear system Algorithm: Newton’s method
19
any step with and ensures descent Nonlinear equations Goal: solve a nonlinear system Algorithm: Newton’s method (Eisenstat and Walker, 1994) (Dembo, Eisenstat, and Steihaug, 1982)
20
Line Search SQP Framework Define “exact” penalty function Implement a line search
21
Exact Case
22
Exact step minimizes the objective on the linearized constraints
23
Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective
24
Quadratic/linear model of merit function Create model Quantify reduction obtained from step
25
Quadratic/linear model of merit function Create model Quantify reduction obtained from step
26
Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective
27
Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective … but this is ok since we can account for this conflict by increasing the penalty parameter
28
Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective … but this is ok since we can account for this conflict by increasing the penalty parameter
29
for k = 0, 1, 2, … Compute step by… Set penalty parameter to ensure descent on… Perform backtracking line search to satisfy… Update iterate Algorithm Outline (exact steps)
30
First attempt Proposition: sufficiently small residual 1e-81e-71e-61e-51e-41e-31e-21e-1 Success100% 97% 90%85%72%38% Failure0% 3% 10%15%28%62% Test: 61 problems from CUTEr test set
31
First attempt… not robust Proposition: sufficiently small residual … not enough for complete robustness We have multiple goals (feasibility and optimality) Lagrange multipliers may be completely off … may not have descent!
32
Recall the line search condition Second attempt Step computation: inexact SQP step We can show
33
Recall the line search condition Second attempt Step computation: inexact SQP step We can show... but how negative should this be?
34
for k = 0, 1, 2, … Compute step Set penalty parameter to ensure descent Perform backtracking line search Update iterate Algorithm Outline (exact steps)
35
for k = 0, 1, 2, … Compute step and set penalty parameter to ensure descent and a stable algorithm Perform backtracking line search Update iterate Algorithm Outline (inexact steps)
36
Inexact Case
38
Step is acceptable if for
39
Inexact Case Step is acceptable if for
40
Inexact Case Step is acceptable if for
41
for k = 0, 1, 2, … Iteratively solve Until Update penalty parameter Perform backtracking line search Update iterate Algorithm Outline or
42
Observe KKT conditions Termination Test
43
Outline Problem Formulation Equality constrained optimization Sequential Quadratic Programming Inexact Framework Unconstrained optimization and nonlinear equations Stopping conditions for linear solver Global Behavior Merit function and sufficient decrease Satisfying first order conditions Numerical Results Model inverse problem Accuracy tradeoffs Final Remarks Future work Negative curvature
44
The sequence of iterates is contained in a convex set and the following conditions hold: the objective and constraint functions and their first and second derivatives are bounded the multiplier estimates are bounded the constraint Jacobians have full row rank and their smallest singular values are bounded below by a positive constant the Hessian of the Lagrangian is positive definite with smallest eigenvalue bounded below by a positive constant Assumptions
45
Sufficient Reduction to Sufficient Decrease Taylor expansion of merit function yields Accepted step satisfies
46
Intermediate Results is bounded below by a positive constant is bounded above
47
Sufficient Decrease in Merit Function
48
Step in Dual Space (for sufficiently small and ) Therefore, We converge to an optimal primal solution, and
49
Outline Problem Formulation Equality constrained optimization Sequential Quadratic Programming Inexact Framework Unconstrained optimization and nonlinear equations Stopping conditions for linear solver Global Behavior Merit function and sufficient decrease Satisfying first order conditions Numerical Results Model inverse problem Accuracy tradeoffs Final Remarks Future work Negative curvature
50
Problem Formulation Tikhonov-style regularized inverse problem Want to solve for a reasonably large mesh size Want to solve for small regularization parameter SymQMR for linear system solves Input parameters: orRecall: (Curtis and Haber, 2007)
51
Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 0.52929.5s145250.13.12e-1 0.11211.37s65454.56.90e-2 0.01911.60s68175.76.27e-3 n1024 m512 1e-6 (Curtis and Haber, 2007)
52
Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 0.52929.5s145250.13.12e-1 0.11211.37s65454.56.90e-2 0.01911.60s68175.76.27e-3 n1024 m512 1e-6 (Curtis and Haber, 2007)
53
Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 1e-61211.40s65454.56.90e-2 1e-71114.52s84076.46.99e-2 1e-8810.57s63979.96.15e-2 1e-91118.52s11391048.65e-2 1e-101944.41s27081438.90e-2 n1024 m512 1e-1 (Curtis and Haber, 2007)
54
Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 1e-615264.47s19921338.13e-2 1e-711236.51s17761616.89e-2 1e-89204.51s15671746.77e-2 1e-911347.66s26812448.29e-2 1e-1016805.14s62493918.93e-2 n8192 m4096 1e-1 (Curtis and Haber, 2007)
55
Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 1e-6155055.9s43652918.46e-2 1e-7104202.6s36303638.87e-2 1e-8125686.2s48254027.96e-2 1e-9126678.7s56334698.77e-2 1e-101414783s125258958.63e-2 n65536 m32768 1e-1 (Curtis and Haber, 2007)
56
Outline Problem Formulation Equality constrained optimization Sequential Quadratic Programming Inexact Framework Unconstrained optimization and nonlinear equations Stopping conditions for linear solver Global Behavior Merit function and sufficient decrease Satisfying first order conditions Numerical Results Model inverse problem Accuracy tradeoffs Final Remarks Future work Negative curvature
57
Review and Future Challenges Review Defined a globally convergent inexact SQP algorithm Require only inexact solutions of primal-dual system Require only matrix-vector products involving objective and constraint function derivatives Results also apply when only reduced Hessian of Lagrangian is assumed to be positive definite Numerical experience on model problem is promising Future challenges (Nearly) Singular constraint Jacobians Inexact derivative information Negative curvature etc., etc., etc….
58
Negative Curvature Big question What is the best way to handle negative curvature (i.e., when the reduced Hessian may be indefinite)? Small question What is the best way to handle negative curvature in the context of our inexact SQP algorithm? We have no inertia information! Smaller question When can we handle negative curvature in the context of our inexact SQP algorithm with NO algorithmic modifications? When do we know that a given step is OK? Our analysis of the inexact case leads to a few observations…
59
Why Quadratic Models?
60
Provides a good… direction? Yes step length? Yes Provides a good… direction? Maybe step length? Maybe
61
Why Quadratic Models? One can use our stopping criteria as a mechanism for determining which are good directions All that needs to be determined is whether the step lengths are acceptable
62
Unconstrained Optimization Direct method is the angle test Indirect method is to check the conditions or
63
Unconstrained Optimization Direct method is the angle test Indirect method is to check the conditions or step qualitystep length
64
Constrained Optimization Step quality determined by Step length determined by or
65
Thanks!
66
Actual Stopping Criteria Stopping conditions: Model reduction condition or
67
Constraint Feasible Case If feasible, conditions reduce to
68
Constraint Feasible Case If feasible, conditions reduce to
69
Constraint Feasible Case If feasible, conditions reduce to Some region around the exact solution
70
Constraint Feasible Case If feasible, conditions reduce to Ellipse distorted toward the linearized constraints
71
Constraint Feasible Case If feasible, conditions reduce to
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.