CS5321 Numerical Optimization 18 Sequential Quadratic Programming (Active Set methods) 12/6/2018
Local SQP model The problem minxf (x) subject to c(x)=0 can be modeled as a quadratic programming at x=xk Assumptions: A(x), the constraint Jacobian, has full row rank. xx2L(x,) is positive definite on the tangent space of constraints, that is, dTxx2L(x,)d>0 for all d0,Ad=0. m i n p k ( ) = f + r T 1 2 x L s . t A c 12/6/2018
Inequality constraints For The local quadratic model is m i n x f ( ) s . t c = ; 2 E ¸ I m i n p k ( ) = f + r T 1 2 x L s . t c ; E ¸ I 12/6/2018
Theorem of active set methods Theorem 18.1 (Robinson 1974) If x* is a local solution of the original problem with some *, and the pair (x*, *) satisfies the KKT condition, the LICO condition, and the second order sufficient conditions, then for (xk, k) sufficiently close to (x*, *), then there is a local quadratic model whose active set is the same as that of the original problem. 12/6/2018
Sequential QP method Choose an initial guess x0, 0 For k = 1, 2, 3, … Evaluate fk, fk, xx2Lk, ck, and ck(=Ak) Solve the local quadratic programming Set xk+1 = xk + pk How to choose the active set? How to solve 2(b)? Haven’t we solved that in chap 16? (Yes and No) 12/6/2018
Algorithms Two types of algorithms to choose active set Inequality constrained QP (IQP): Solve QPs with inequality constraints and take the local active set as the optimal one. Equality constrained QP (EQP): Select constraints as the active set and solve equality constrained QPs. Basic algorithms to solve 2(b) Line search methods Trust region methods Nonlinear gradient projection 12/6/2018
Solving SQP All techniques in chap 16 can be applied. But there are additional problems need be solved Linearized constraints may not be consistent Convergence guarantee (Hessian is not positive def) And some useful properties can be used Hessian can be updated by the quasi-Newton method Solutions can be used as an initial guess (warm start) The exact solution is not required. 12/6/2018
Inconsistent linearizations Linearizing at xk=1 gives The constraints cannot be enforced since they may not exact or consistent. Use penalty function 1 ¡ x ¸ 2 4 ¡ p ¸ 2 3 m i n x ; v w t f ( ) + ¹ X 2 E I s u b j e c o = ¸ ¡ 12/6/2018
Quasi-Newton approximations k = x + 1 ¡ y r L ( ; ¸ ) Recall for the update of Hessian is (BFGS, chap 6) If the updated Hessian Bk+1 is not positive definite Condition fails. Define for (0,1) B k + 1 = ¡ s T y s T k B + 1 = y > r k = µ y + ( 1 ¡ ) B s B k + 1 = ¡ s T r 12/6/2018
BFGS for reduced-Hessian T = ¡ Q 1 2 ¢ µ R ¶ ; p + Let Need to solve Solve * first to obtained an active set. Ignore term. Solve The reduced secant formula is Use BFGS on this equation. R p 1 = ¡ h ( Q T 2 G ) g ¸ ¤ + Q T 2 G 1 p ( Q T 2 G ) p = ¡ g ( Q T 2 G ) k + 1 ® p = [ r x L ; ¸ ¡ ] 12/6/2018
1.Line search SQP method Set step length k such that the merit function is sufficiently decreased? One of the Wolfe condition (chap 3) D(1, pk) is the directional derivative of 1 in pk. (theorem 18.2 ) Let k =1 and decrease it until the condition is satisfied. Á 1 ( x ; ¹ ) = f + k c Á 1 ( x k + ® p ; ¹ ) · ´ D D ( Á 1 x k ; ¹ ) p = r f T ¡ c 12/6/2018
2.Trust region SQP method Problem modification Relax the constraints (rk=0 in the original algorithm) Add trust region radius as a constraint There are smart ways to choose rk. For example, m i n p k ( ) = f + r T 1 2 x L s . t A c ; · ¢ m i n v r k ( ) = A + c 2 s . t · : 8 ¢ 12/6/2018
3.Nonlinear gradient projection Let Bk be the s.p.d. approximation to 2f (xk). Step direction is pk=x−xk Combine with the line-search dir xk+1=xk+kpk. Choose k s.t. f (xk+1) f (xk)+kfkTpk. Combine with the trust region bounds ||pk||k. m i n x q k ( ) = f + r T ¡ 1 2 B s . t l · u m i n x q k ( ) = f + r T ¡ 1 2 B s . t a l ; ¢ e · u 12/6/2018