Download presentation
Presentation is loading. Please wait.
Published byRoland Beasley Modified over 6 years ago
1
Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi
LP, SOCP, and Optimization-Free Approaches to Polynomial Optimization and Difference of Convex Programming Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi
2
Two problems of interest
Lower bounds for polynomial optimization problems Writing a polynomial as a difference of two convex polynomials Can we provide lower bounds on the polynomial optimization problem (POP) min π₯β β π π(π₯) s.t. π π π₯ β₯0, π=1,β¦,π where π, π π , π=1,β¦,π are polynomials? Given a polynomial π π₯ , can we find convex polynomials π π₯ and β(π₯) such that π π₯ =π π₯ ββ π₯ ? First part of the talk Second part of the talk
3
Part I: Computing lower bounds for POPs
4
Lower bounds on POPs: applications (1/2)
Polynomial optimization problems min π₯β β π π(π₯) π .π‘. π π π₯ β₯0 Optimal power flow problem Economics and game theory Sensor network localization
5
Lower bounds on POPs: applications (2/2)
Infeasibility of a set of polynomial inequalities The system π π π₯ β₯0, π=1,β¦,π is infeasible. 0 is a strict lower bound on min π₯ βπ 1 (π₯) s.t. π 2 π₯ β₯0,β¦, π π π₯ β₯0 β If 0 is not a strict upper bound on g_1(x) then it means that there exists x such that g2(x)>=0,β¦,gm(x)>=0 and g1(x)>=0 which means the system is feasible Automated theorem proving Optimality certificates for discrete optimization
6
Lower bounds on POPs: the SOS approach (1/2)
A polynomial π is a sum of squares (SOS) if it can be written as π π₯ = π π π 2 (π₯) , where π π are polynomials. Example: Sufficient condition for nonnegativity. A polynomial π(π₯) of degree 2d is SOS if and only if βπβ½0 such that where π§= 1, π₯ 1 ,β¦, π₯ π , π₯ 1 π₯ 2 ,β¦, π₯ π π π is the vector of monomials of degree up to π.
7
Lower bounds on POPs: the SOS approach (2/2)
max πΎ πΎ π .π‘. π π₯ βπΎβ₯0, βπ₯β π₯ π π π₯ β₯0, π=1,β¦,π} min π₯ π(π₯) π .π‘. π π π₯ β₯0, π=1,β¦,π = opt. val. πΈ β Under compactness assumptions, for large enough degree = opt. val. max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ + π π π π₯ β
π π π₯ , where π π , π=0,β¦,π, are SOS. For fixed degrees of the SOS polynomials, solving this problem is an SDP This is an SOS certificate that πΎ is a lower bound for POP. There are many other such SOS certificates.
8
The first part of the talk
Goal: avoid using SOS polynomials and SDP to compute lower bounds for POPs LP and SOCP-based alternatives to SOS polynomials (applies to any SOS-based certificate) Using dsos/sdsos polynomials and improvements New certificate that πΎ is a lower bound on the POP that only requires checking if coefficients of a given polynomial are β₯0 Providing a new certificate that does not use optimization at all max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ + π π π π₯ β
π π π₯ , where π π , π=0,β¦,π, are SOS. max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ + π π π π₯ β
π π π₯ , where π π , π=0,β¦,π, are SOS. dsos/sdsos + improvements New certificate
9
Part I: Computing lower bounds for POPs
Using dsos/sdsos polynomials and improvements Providing an optimization-free certificate for lower bounds on POPs
10
LP and SOCP alternatives to sos: dsos and sdsos
Sum of squares (sos) π π₯ =π§ π₯ π ππ§ π₯ , πβ½0 SDP DD cone β πΈ πΈ ππ β₯ π πΈ ππ , βπ} PSD coneβ πΈ πΈβ½π} SDD cone β πΈ β diagonal π« with π« ππ >π s.t. π«πΈπ« π
π
} Diagonally dominant sum of squares (dsos) π π₯ =π§ π₯ π ππ§ π₯ , π ππππππππππ¦ ππππππππ‘ (dd) LP Scaled diagonally dominant sum of squares (sdsos) π π₯ =π§ π₯ π ππ§ π₯ , π π πππππ ππππππππππ¦ ππππππππ‘ (sdd) SOCP Ahmadi, Majumdar
11
where π π , π=0,β¦,π, are dsos/sdsos.
LP and SOCP alternatives to sos: dsos and sdsos Problem: computing lower bounds for POPs max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ + π π π π₯ β
π π π₯ , where π π , π=0,β¦,π, are sos. max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ + π π π π₯ β
π π π₯ , where π π , π=0,β¦,π, are dsos/sdsos. β₯ opt. val. b Example: For a parametric family of polynomials: π π₯ 1 , π₯ 2 =2 π₯ π₯ 2 4 +π π₯ 1 3 π₯ 2 +(1βπ) π₯ 1 2 π₯ 2 2 +π π₯ 1 π₯ 2 3 a
12
Improvements on dsos and sdsos (1/2)
Replacing sos polynomials by dsos/sdsos polynomials: +: fast bounds - : not always as good quality (compared to sos) Iteratively construct a sequence of improving LP/SOCPs Initialization: Start with the dsos/sdsos polynomials Column Generation method
13
Improvements on dsos/sdsos (2/2)
Problem of interest: computing lower bounds for POPs Example considered : computing lower bounds for unconstrained POPs Compactness assumptions + large degree max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ + π π π π₯ β
π π π₯ , where π π , π=0,β¦,π, are SOS. min π₯ π(π₯) π .π‘. π π π₯ β₯0 = opt. val. max πΎ πΎ π .π‘. π π₯ βπΎ= π 0 π₯ , where π 0 is SOS. β₯ opt. val. min π₯ π(π₯)
14
Method 2: Column generation (1/3)
Focus on LP-based version of this method (SOCP is similar). Sos problem max πΎ πΎ π .π‘. π π₯ βπΎ SOS Replace with Dsos problem max πΎ πΎ π .π‘. π π₯ βπΎ dsos Dsos problem max πΎ πΎ π .π‘. π π₯ βπΎ=π π π» πΈπ π , πΈ π
π
Two different ways of characterizing πΈ π
π
: β‘ π dd ββ πΌ π β₯0 s.t. π= π πΌ π π£ π π£ π π , where π£ π fixed vector with at most two nonzero components =Β±1 β π dd β π ππ β₯ π |π ππ | , β π [In collaboration with Sanjeeb Dash, IBM Research]
15
Method 2: Column generation (2/3)
Dsos problem max πΎ πΎ π .π‘. π π₯ βπΎ= π πΆ π π π π» π π π , πΆ π β₯π Dsos problem max πΎ πΎ π .π‘. π π₯ βπΎ=π π π» πΈπ π , πΈ π
π
Using β‘ Idea behind the algorithm: Expand feasible space at each iteration by adding a new vector π£ and variable πΌ max πΎ πΎ π π βπΈ= π πΆ π π π π» π π π +πΆ π π» π π π πΆ π β₯π, πβ₯π Question: How to pick π? Use the dual!
16
Method 2: Column Generation (3/3)
DSOS + 5 iterations SDSOS + 5 iterations Example: minimizing a degree-4 polynomial π=ππ π=ππ π=ππ π=ππ π=ππ bd t(s) DSOS -10.96 0.38 -18.01 0.74 -26.45 15.51 -36.52 7.88 -62.30 10.68 π·ππ π π -5.57 31.19 -9.02 471.39 -20.08 600 -32.28 -35.14 SOS -3.26 5.60 -3.58 82.22 -3.71 NA
17
Part I: Avoiding SDP to compute lower bounds on POPs
Using dsos/sdsos polynomials and improvements Providing an optimization-free certificate for lower bounds on POPs
18
An optimization free method for lower bounds (1/3)
max πΎ πΎ π .π‘. π π₯ βπΎβ₯0, βπ₯β π₯ π π π₯ β₯0, π=1,β¦,π} min π₯β β π π(π₯) π .π‘. π π π₯ β₯0, π=1,β¦,π = opt. val. 2π= maximum degree of π, π π Under compactness assumptions, i.e., π₯ π π π₯ β₯0}βπ΅(0,π
) = opt. val. For large enough π sup πΎ πΎ s.t. π πΎ π£ 2 β π€ 2 β 1 π π (π£ π 2 β π€ π 2 ) 2 π + 1 2π π π£ π 4 + π€ π π β
π π£ π 2 + π π€ π π 2 has nonnegative coefficients, where π πΎ is a form in π+π+3 variables and of degree 4π which explicitely depends on π, π π and π
Theorem:
19
An optimization free method for lower bounds (2/3)
πΎ is a strict lower bound on the POP β βπββ s.t. π πΈ π π β π π β π π π (π π π β π π π ) π π
+ π ππ π π π π + π π π π
β
π π£ π 2 + π π€ π π 2 has β₯0 coefficients Proof sketch of the theorem: πΎ is a strict lower bound on the POP β βπββ s.t. π πΈ π π β π π β 1 π π (π£ π 2 β π€ π 2 ) 2 π + π ππ π π π π + π π π π
β
π π£ π 2 + π π€ π π 2 has β₯0 coefficients πΎ is a strict lower bound on the POP β βπββ s.t. π πΈ π π β π π β 1 π π (π£ π 2 β π€ π 2 ) 2 π + 1 2π π π£ π 4 + π€ π π β
π π£ π 2 + π π€ π π 2 has β₯0 coefficients πΎ is a strict lower bound on the POP β βπββ s.t. π πΎ π£ 2 β π€ 2 β 1 π π (π£ π 2 β π€ π 2 ) 2 π + 1 2π π π£ π 4 + π€ π π β
π π£ π 2 + π π€ π π 2 has β₯0 coefficients πΎ is a lower bound on the POP β π πΎ is pd Result by Polya: π even and pd ββ πββ such that π π§ β
π π§ π 2 π has nonnegative coefficients. Make π πΎ (π§) even by considering π πΈ π π β π π . We lose positive definiteness of π πΎ with this transformation. Add the positive definite term π ππ π π π π + π π π ) to π πΎ ( π£ 2 β π€ 2 ) to make it positive definite. Apply Polyaβs result. The term β π π π (π π π β π π π ) π π
ensures that the converse holds as well.
20
An optimization free method for lower bounds (3/3)
How to use this method in practice? Given a polynomial optimization problem: min π(π₯) , s.t. π π π₯ β₯0, π
such that π₯ π π π₯ β₯0}βπ΅(0,π
), an accuracy π for the optimal value of the POP π β Expand and check nonnegativity of the coefficients of the polynomial π πΎ π£ 2 β π€ 2 β 1 π π (π£ π 2 β π€ π 2 ) 2 π + 1 2π π π£ π 4 + π€ π π β
π π£ π 2 + π π€ π π 2 Have we reached accuracy π in bisection? Start bisection on πΎ by taking πΎ= π 0 + πΏ 0 2 π=1 NB: πΏ 0 β€ π β β€ π 0 No: keep doing bisection Yes β¦ π=2
21
Part II: Writing a polynomial as a difference of two convex polynomials
22
Difference of convex decomposition: Applications
Can be used for difference of convex programs, i.e., problems of the form: min π 0 (π₯) π .π‘. π π π₯ β€0 where π π π₯ β π π π₯ β β π π₯ , π π , β π convex. Applications: Machine Learning (Sparse PCA, Kernel selection, feature selection in SVM) Studied for quadratics, polynomials nice to study question computationally Hiriart-Urruty, 1985 Tuy, 1995
23
Difference of Convex (dc) decomposition
Difference of convex (dc) decomposition: given a polynomial π, find π and β such that π=πβπ, where π,β convex polynomials. Questions: Does such a decomposition always exist? Can I obtain such a decomposition efficiently? Is this decomposition unique?
24
Existence of dc decomposition (1/5)
Recall: Theorem: Any polynomial can be written as the difference of two dsos-convex polynomials. Corollary: Any polynomial can be written as the difference of two sdsos-convex/sos-convex/convex polynomials. π¦ π π» π π₯ π¦ sos SDP SOS-convexity π¦ π π» π π₯ π¦ sdsos SOCP SDSOS-convexity π¦ π π» π π₯ π¦ dsos LP DSOS-convexity π(π₯) convex π¦ π π» π π₯ π¦β₯0, βπ₯,π¦β β π β β β
25
Existence of dc decomposition (2/5)
Lemma: Let πΎ be a full dimensional cone in a vector space πΈ. Then any π£β πΈ can be written as π£= π 1 β π 2 , π 1 , π 2 βπΎ. Proof sketch: =:πβ² β πΌ<1 such that 1βπΌ π£+πΌπβπΎ E K βπ£= 1 1βπΌ π β² β πΌ 1βπΌ π To change π πβ² π π 1 βπΎ π 2 βπΎ
26
Existence of dc decomposition (3/5)
Here, πΈ={polynomials of degree d, in n variables}, πΎ={dsos-convex polynomials of degree d and in n variables }. Remains to show that πΎ is full dimensional: the polynomial in the interior is constructed inductively on π with appropriately scaled coefficients. This shows that a decomposition can be obtained efficiently: π=πβπ, π,π s/d/sos-convex solving is an LP/SOCP/SDP.
27
Uniqueness of dc decomposition
Dc decomposition: given a polynomial π, find convex polynomials π and β such that π=πβπ. Questions: Does such a decomposition always exist? Can I obtain such a decomposition efficiently? Is this decomposition unique? οΌ Yes οΌ Through s/d/sos-convexity Alternative decompositions π π₯ = π π₯ +π π₯ β β π₯ +π π₯ π(π₯) convex Initial decomposition xπ π₯ =π π₯ ββ(π₯) βBest decomposition?β
28
Convex-Concave Procedure (CCP)
Heuristic for minimizing DC programming problems. Idea: Input πβ0 x π₯ 0 , initial point π π = π π β β π , π=0,β¦,π Convexify by linearizing π x π π π π = π π π₯ β( β π π₯ π +π» β π π₯ π π π₯β π₯ π ) Solve convex subproblem Take π₯ π+1 to be the solution of min π 0 π π₯ π .π‘. π π π π₯ β€0, π=1,β¦,π convex convex affine πβπ+1 π π π π π π (π)
29
Convex-Concave Procedure (CCP)
Toy example: min π₯ π π₯ , where π π₯ βπ π₯ ββ(π₯) Convexify π π₯ to obtain π 0 (π₯) Initial point: π₯ 0 =2 Minimize π 0 (π₯) and obtain π₯ 1 Reiterate π₯ β π₯ 3 π₯ 4 π₯ 2 π₯ 1 π₯ 0 π₯ 0
30
Picking the βbestβ decomposition for CCP
Algorithm Linearize π π around a point π₯ π to obtain convexified version of π(π) Algorithm Linearize π π around a point π₯ π to obtain convexified version of π(π) Algorithm Linearize π π around a point π₯ π to obtain convexified version of π(π) Idea Pick β π₯ such that it is as close as possible to affine around π₯ π Idea Pick β π₯ such that it is as close as possible to affine around π₯ π Mathematical translation Minimize curvature of β
31
Undominated decompositions (1/2)
Definition: g ,ββπβf is an undominated decomposition of π if no other decomposition of π can be obtained by subtracting a (nonaffine) convex function from π. π π = π π + π π , π π =π π π +ππβπ Convexify around π₯ 0 =2 to get π π π π π = π π βπ π π +ππβπ Cannot substract something convex from g and get something convex again. DOMINATED BY π β² π = π π , π β² π =π π π +ππβπ Convexify around π₯ 0 =2 to get π πβ² π If πβ² dominates π then the next iterate in CCP obtained using π β² always beats the one obtained using π.
32
Undominated decompositions (2/2)
Theorem: Given a polynomial π, consider min 1 π΄ π π πβ1 ππ π» π ππ , (where π΄ π = 2 π π/2 Ξ(π/2) ) s.t. π=πββ, π convex, β convex Any optimal solution is an undominated dcd of π (and an optimal solution always exists). Theorem: If π has degree 4, it is strongly NP-hard to solve (β). Idea: Replace π=πββ, π, β convex by π=πββ, π,β sos-convex. (β) π,β
33
Comparing different decompositions (1/2)
Solving the problem: min π΅= π₯ π₯ β€π
} π 0 , where π 0 has π=8 and π=4. Decompose π 0 , run CCP for 4 minutes and compare objective value. Feasibility Undominated min π,β 1 π΄ π π πβ1 ππ π» π ππ π .π‘. π 0 =πββ π,β sos-convex min g,h 0 s.t. π 0 =πββ π,β sos-convex
34
Comparing different decompositions (2/2)
Average over 30 instances Solver: Mosek Computer: 8Gb RAM, 2.40GHz processor Feasibility Undominated Conclusion: Performance of CCP strongly affected by initial decomposition.
35
Main messages (1/2) Computing lower bounds on POPs has many applications. Traditionally, one can obtain lower bounds on POPs using sos polynomials. Here, we presented two different methods to not use sos polynomials. Replace sos polynomials by dsos/sdsos polynomials Improve on these via sequences of LPs and SOCPs Use a new optimization-free certificate to certify lower bounds on POPs (only uses multiplication and coefficient checking)
36
Main messages (2/2) Decomposing a polynomial into the difference of two convex polynomials has long been a problem of interest. A decomposition can be obtained efficiently using sos-convexity and SDP We can also obtain decompositions using LP and SOCP. We show that the decomposition we pick impacts performance of the CCP algorithm, a heuristic to solve DC programs.
37
Thank you for listening
Questions? Want to learn more?
38
Method 1: Cholesky change of basis (1/3)
dd in the βright basisβ psd but not dd Goal: iteratively improve on basis
39
Method 1: Cholesky change of basis (2/3)
Sos problem max πΎ πΎ π .π‘. π π₯ βπΎ SOS Initialize max πΎ πΎ s.t. π π₯ βπΎ=π π π» πΈπ π , πΈ π
π
/ππ
π
Step 1 Replace: π 1 =πβππ( π β ) Step 1 Replace: π π =πβππ( π πβ1 π π β π πβ1 ) Step 2 max πΎ πΎ s.t. π π₯ βπΎ=π π π» πΌ π π» πΈ πΌ π π π , πΈ π
π
/ππ
π
Step 2 max πΎ πΎ π .π‘. π π₯ βπΎ=π π π» πΌ π π» πΈ πΌ π π π , πΈ π
π
/ππ
π
New basis πβπ+π One iteration of this method on a parametric family of polynomials: π π₯ 1 , π₯ 2 =2 π₯ π₯ 2 4 +π π₯ 1 3 π₯ 2 +(1βπ) π₯ 1 2 π₯ 2 2 +π π₯ 1 π₯ 2 3
40
Method 1: Cholesky change of basis (3/3)
Example: min π₯ π(π₯) , where π is in 4 variables and has degree 4 Lower bound on optimal value
41
Method 2: Column generation (3/4)
PRIMAL DUAL A general SDP max π¦β β π π π π¦ π .π‘. πΆβ π=1 π π¦ π π΄ π β½0 πΊπ«π· Dual of SDP min πβ π π π‘π(πΆπ) π .π‘. π‘π π΄ π π = π π πβ½0 π³π· πΊπ« π· β π³ π· β LP obtained with inner approximation of PSD by DD Dual of LP min πβ π π π‘π(πΆπ) π .π‘. π‘π π΄ π π = π π π£ π π π π£ π β₯0 max π¦β β π π π π¦ π .π‘. πΆβ π=1 π π¦ π π΄ π = πΌ π π£ π π£ π π πΌ π β₯0 Pick π s.t. π π» πΏπ<π.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.