1 On the Cost and Benefits of Procrastination: Approximation Algorithms for Stochastic Combinatorial Optimization Problems (SODA 2004) Nicole Immorlica,

Slides:



Advertisements
Similar presentations
Iterative Rounding and Iterative Relaxation
Advertisements

The Primal-Dual Method: Steiner Forest TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A AA A A A AA A A.
Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
Primal-Dual Algorithms for Connected Facility Location Chaitanya SwamyAmit Kumar Cornell University.
1 Routing and Wavelength Assignment in Wavelength Routing Networks.
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
Combinatorial Algorithms
Approximation Some Network Design Problems With Node Costs Guy Kortsarz Rutgers University, Camden, NJ Joint work with Zeev Nutov The Open University,
24 Feb 04Boosted Sampling1 Boosted Sampling: Approximation Algorithms for Stochastic Problems Martin Pál Joint work with Anupam Gupta R. Ravi Amitabh Sinha.
1 Online and Stochastic Survivable Network Design Ravishankar Krishnaswamy Carnegie Mellon University joint work with Anupam Gupta and R. Ravi.
1 Stochastic Event Capture Using Mobile Sensors Subject to a Quality Metric Nabhendra Bisnik, Alhussein A. Abouzeid, and Volkan Isler Rensselaer Polytechnic.
Approximation Algorithms
The Complexity of the Network Design Problem Networks, 1978 Classic Paper Reading
1 Approximation Algorithms for Demand- Robust and Stochastic Min-Cut Problems Vineet Goyal Carnegie Mellon University Based on, [Golovin, G, Ravi] (STACS’06)
A Constant Factor Approximation Algorithm for the Multicommodity Rent-or-Buy Problem Amit Kumar Anupam Gupta Tim Roughgarden Bell Labs CMU Cornell joint.
Primal Dual Method Lecture 20: March 28 primaldual restricted primal restricted dual y z found x, succeed! Construct a better dual.
Approximation Algorithm: Iterative Rounding Lecture 15: March 9.
Implicit Hitting Set Problems Richard M. Karp Harvard University August 29, 2011.
A general approximation technique for constrained forest problems Michael X. Goemans & David P. Williamson Presented by: Yonatan Elhanani & Yuval Cohen.
Sublinear Algorithms for Approximating Graph Parameters Dana Ron Tel-Aviv University.
Approximation Algorithms
Robust Network Design with Exponential Scenarios By: Rohit Khandekar Guy Kortsarz Vahab Mirrokni Mohammad Salavatipour.
An Approximation Algorithm for Requirement cut on graphs Viswanath Nagarajan Joint work with R. Ravi.
Near-Optimal Network Design with Selfish Agents By Elliot Anshelevich, Anirban Dasgupta, Eva Tardos, Tom Wexler STOC’03 Presented by Mustafa Suleyman CIFTCI.
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
On Stochastic Minimum Spanning Trees Kedar Dhamdhere Computer Science Department Joint work with: Mohit Singh, R. Ravi (IPCO 05)
Network Design with Concave Cost Functions Kamesh Munagala, Stanford University.
Primal-Dual Algorithms for Connected Facility Location Chaitanya SwamyAmit Kumar Cornell University.
1 Introduction to Approximation Algorithms Lecture 15: Mar 5.
A General Approach to Online Network Optimization Problems Seffi Naor Computer Science Dept. Technion Haifa, Israel Joint work: Noga Alon, Yossi Azar,
Packing Element-Disjoint Steiner Trees Mohammad R. Salavatipour Department of Computing Science University of Alberta Joint with Joseph Cheriyan Department.
Approximation Algorithms: Bristol Summer School 2008 Seffi Naor Computer Science Dept. Technion Haifa, Israel TexPoint fonts used in EMF. Read the TexPoint.
TECH Computer Science Graph Optimization Problems and Greedy Algorithms Greedy Algorithms  // Make the best choice now! Optimization Problems  Minimizing.
Sampling-based Approximation Algorithms for Multi-stage Stochastic Optimization Chaitanya Swamy Caltech and U. Waterloo Joint work with David Shmoys Cornell.
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
ICALP'05Stochastic Steiner without a Root1 Stochastic Steiner Trees without a Root Martin Pál Joint work with Anupam Gupta.
Approximation Algorithms for Stochastic Combinatorial Optimization Part I: Multistage problems Anupam Gupta Carnegie Mellon University.
V. V. Vazirani. Approximation Algorithms Chapters 3 & 22
The Multiplicative Weights Update Method Based on Arora, Hazan & Kale (2005) Mashor Housh Oded Cats Advanced simulation methods Prof. Rubinstein.
Design Techniques for Approximation Algorithms and Approximation Classes.
Approximating Minimum Bounded Degree Spanning Tree (MBDST) Mohit Singh and Lap Chi Lau “Approximating Minimum Bounded DegreeApproximating Minimum Bounded.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Topics in Algorithms 2005 Constructing Well-Connected Networks via Linear Programming and Primal Dual Algorithms Ramesh Hariharan.
Approximation Algorithms Department of Mathematics and Computer Science Drexel University.
Edge-disjoint induced subgraphs with given minimum degree Raphael Yuster 2012.
Primal-Dual Algorithms for Connected Facility Location Chaitanya SwamyAmit Kumar Cornell University.
Approximation Algorithms
Approximation Algorithms for Stochastic Optimization Chaitanya Swamy Caltech and U. Waterloo Joint work with David Shmoys Cornell University.
Stochastic Multicast with Network Coding Ajay Gopinathan, Zongpeng Li Department of Computer Science University of Calgary ICDCS 2009, June , Montreal.
Models in I.E. Lectures Introduction to Optimization Models: Shortest Paths.
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
Unit 9: Coping with NP-Completeness
1 Mar 05Stochastic Steiner without a Root1 Stochastic Steiner Tree without a Root Martin Pál Joint work with Anupam Gupta.
Multicommodity flow, well-linked terminals and routing problems Chandra Chekuri Lucent Bell Labs Joint work with Sanjeev Khanna and Bruce Shepherd Mostly.
CS223 Advanced Data Structures and Algorithms 1 Maximum Flow Neil Tang 3/30/2010.
Approximation Algorithms Department of Mathematics and Computer Science Drexel University.
Implicit Hitting Set Problems Richard M. Karp Erick Moreno Centeno DIMACS 20 th Anniversary.
Lecture.6. Table of Contents Lp –rounding Dual Fitting LP-Duality.
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
TU/e Algorithms (2IL15) – Lecture 12 1 Linear Programming.
Approximation Algorithms based on linear programming.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
The Theory of NP-Completeness
6.5 Stochastic Prog. and Benders’ decomposition
1.3 Modeling with exponentially many constr.
1.3 Modeling with exponentially many constr.
Integer Programming (정수계획법)
The Theory of NP-Completeness
6.5 Stochastic Prog. and Benders’ decomposition
Chapter 1. Formulations.
Presentation transcript:

1 On the Cost and Benefits of Procrastination: Approximation Algorithms for Stochastic Combinatorial Optimization Problems (SODA 2004) Nicole Immorlica, David Karger, Maria Minkoff, Vahab S. Mirrokni Jaehan Koh Dept. of Computer Science Texas A&M University

2 Outline Introduction Preplanning Framework Examples Results Summary Homework

3 Introduction Scenarios Harry is designing a network for the Computer Science Department at Texas A&M University. In designing, he should make his best guess about the future demands in a network and purchase capacity in accordance with them. A mobile robot is navigating around a room. Since the information about the environment is unknown or becomes available too late to be useful to the robot, it might be impossible to modify a solution or improve its value once the actual inputs are revealed.

4 Planning under uncertainty Problem data frequently subject to uncertainty May represent information about the future Inputs may be evolve over time On-line model Assumes no knowledge of the future Stochastic modeling of uncertainty Given: probability distribution on potential outcomes Goal: minimize expected cost over all potential outcomes

5 Approaches Plan ahead Full solution has to be specified before we learn values of unknown parameters Information becomes available too late to be useful Wait-and-See Possibility to defer some decisions until getting to know exact inputs Trade-off: decisions made late may be more expensive

6 Approaches (Cont’d) Trade-offs Make some purchase/allocation decisions early to reduce cost while deferring others at greater expense to take advantage of additional information. Problems in which the probolem instance is uncertain. Min-cost flow, bin packing, vertex cover, shortest path, and the Steiner tree problems.

7 Preplanning framework Stochastic combinatorial optimization problem ground set of elements e  E A (randomly selected) problem instance I, which defines a set of feasible solutions F I, each corresponding to a subset F I  2 E. We can buy certain elements “in advance” at cost c e, then sample a problem instance, then buy other elements at “last-minute” cost c e so as to produce a feasible solution S  F I for the problem instance. Goal: to choose a subset of elements to buy in advance to minimize the expected total cost.

8 Two types of instance prob. distribution Bounded support distribution Nonzero probability to only a polynomial number of distinct problem instances. Independent distribution Each element / constraint for the problem instance active independently with some probability.

9 Versions Scenario-based Bounded number of possible scenarios Explicit probability distribution over problem instances Independent events model Random instance is defined implicitly by underlying probabilistic process The number of possible scenarios can be exponential in the problem size

10 Problems Min Cost Flow Given a source and sink and a probability distribution on demand, buy some edges in advance and some after sampling (at greater cost) such that the given amount of demand can be routed from source to sink. Bin Packing A collection of item is given, each of which will need to be packed into a bin with some probability. Bins can be purchased in advance at cost 1; after the determination of which items need to be packed, additional bins can be purchased as cost > 1. How many bins should be purchased in advance to minimize the expected total cost?

11 Problems (Cont’d) Vertex Cover A graph is given, along with a probability distribution over sets of edges that may need to be covered. Vertices can be purchased in advance at cost 1; after determination of which edges need to be covered, additional vertices can be purchased at cost. Which vertices should be purchased in advance?

12 Problems (Cont’d) Cheap Path Given a graph and a randomly selected pair of vertices (or one fixed vertex and one random vertex), connect them by a path. We can purchase edge e at cost c e before the pair is known or at cost c e after. We wish to minimize the expected total edge cost.

13 Problems (Cont’d) Steiner Tree A graph is given, along with a probability distribution over sets of terminals that need to be connected by a Steiner tree. Edge e can be purchased at cost c e in advance or at cost c e after the set of terminal is known.

14 Preplanning combinatorial optimization problem A ground set of elements e  E, a probability distribution on instances {I}, a cost function c: E  R, a penalty factor  1 Each instance I has a corresponding set of feasible solutions F I  2 E associated with it. Suppose a set of elements A  E is purchased before sampling the probability distribution. The posterior cost function c A is defined by

15 Preplanning combinatorial optimization problem (Cont’d) The objective of a PCO problme: choose a subset of elements A to be purchased in advance so as to minimize the total expected cost of a feasible solution over a random choice of an instance I.

16 The Threshold Property Theorem 1. An element should be purchased in advance if and only if the probability it is used in the solution for a randomly chosen instance exceeds 1/.

17 Example: Min-cost Flow We wish to provide capacity on a network sufficient to carry a random amount of flow demand D from a source s to a sink t. We have an option to pre-install some amout of capacity in advance at some cost per unit. We rent additional capacity once the demands become known, but at cost a factor of or larger per unit. The sum of capacity in advance and capacity rented must satisfy a given upper bound of total capacity for each edge. Goal: Over a given probability distribution on demands, minimize the expected cost of installing sufficient capacity in the network so that the network satisfies the demand.

18 Example: Min-cost Flow (Cont’d) Suppose Cap(s-a) = Cap(a-t) = Cap(s-b) = Cap(a-b) = Cap(b-t) = 1 = 2 Pr[D=0] = ¼, Pr[D=1] = ¼, Pr[D=2] = ½ a s b t

19 Example: Min-cost Flow (Cont’d) Suppose Cap(s-a) = Cap(a-t) = Cap(s-b) = Cap(a-b) = Cap(b-t) = 1 = 2 Pr[D=0] = 5/12, Pr[D=1] = ¼, Pr[D=2] = 1/3 a s b t

20 Example: Vertex cover Classical vertex cover problem Given: a graph G = (V, E) Output: a subset of vertices such that each edge has at least one endpoint in the set Goal: minimize cardinality of the vertex subset Stochastic vertex cover Random subset of edges is present Vertices picked in advance cost 1 Additional vertices can be purchased at cost >1 Goal: minimize expected cost of a vertex cover Time-information trade-off

21 Our techniques Merger of Linear Programs Easy way to handle scenario-based stochastic problems with LP-formulations Threshold Property Identity elements likely to be needed and buy them in advance Probability Aggregation Cluster probability mass to get nearly deterministic subproblems Justifies buying something in advance

22 Vertex Cover with preplanning Given Graph G = (V, E) Random instance: a subset of edges to be covered Scenario-based: poly number of possible edge sets Independent version: each e  E present independently with probability p. Vertices cost 1 in advance; after edge set is sampled Goal: select a subset A of vertices to buy in advance such that minimize expected cost of a vertex cover

23 Idea 1: LP merger Deterministic problem has LP formulation Write separate LP for each scenario of stochastic problem Take probability-weighted linear combination of objective functions Applications of scenario-based vertex cover VC has LP relaxation Combine LP relaxation of scenarios Round in standard way Result: 4-approximation

24 Idea 2: Threshold property Buy element e in advance  Pr[need e]  1/ Advance cost c e Expected buy-later cost c e Pr[need e] Application of independent-events vertex cover Threshold degree k = 1/ p Vertex with degree k is adjacent to an active edge with probability  1/ Not worth buying vertices with degree < k in advance Idea: purchase a subset of high degree vertices in advance

25 k-matching k-matching is a subset of edges that induces degree  k on each vertex Vertex v  V is tight if its degree in matching is exactly k Can construct a maximal k-matching greedily

26 Algorithm for k-matching Assume  4 Set k = 1 / p Construct some maximal k-matching M k Purchase in advance the set of tight vertices A t Solution cost Prepurchase cost | A t | “Wait-and-See” cost  expected size of minimum vertex cover of active edges not already covered by A t

27 Bounding 2 nd stage cost Claim: once purchased all the tight vertices, it is optimal to buy nothing else in advance Vertices in A t cover all edges not in k-matching, then e has at least one tight endpoint Vertices not in A t have degree < k E[cost] a vertex cover in V \ A t -induced subgraph is at most OPT

28 Bounding prepurchase cost Restrict attention to instnace induced by M k Costs no more than OPT Intuition Vertex of degree k = 1/ p is likely to be adjacent to an active edge Active edges are not likely to be clustered together Can show that a tight vertex is useful for the vertex cover with sufficiently high probability  | A t | = O(OPT)

29 Steiner network preplanning Given Undirected graph G = (V, E) with edge costs c e  0 Probability p i of node i becoming active Penalty  1 Goal Buy subset of edges in advance in order to minimize expected cost of s Steiner tree over active nodes Approaches: probability aggregation Cluster vertices into groups of 1/ probability Purchase in advance edges of an MST over clusters

30 CLUSTER algorithm The Steiner tree for the ultrametric case An assignment of edge weights satisfying ĉ uv  max (ĉ u, ĉ v ). The basic idea To cluster nodes into components, each containing clients of total probability mass  (1/ ). Lemma Algorithm CLUSTER produces an MST with the specified properties.

31 CLUSTER algorithm (Cont’d)

32 CLUSTER algorithm (Cont’d) A hub tree h h <T >T

33 Probability Aggregation pp pp p p p p p p p p pp p pp p pp p p p p p 1/

34 Results ProblemStochastic elements Approx. guarantee Min-cost s-t flowDemand1Via LP Bin packingItemsAsymptotic FPAS Using FPAS for determ. version Vertex coverEdges Scenario-based Indep. Events Shortest pathStart node Start & end O(1) ConnFacLoc Rent-or-Buy Steiner treeTerminals3 O(log n) Ultrametrics General case

35 Summary Stochastic combinatorial optimization problems in a novel “preplanning” framework Study of time-information trade-off in problems with uncertain inputs Algorithms to approximately optimize the choice of what to purchase in advance and what to defer Open questions Combinatorial algorithms for stochastic min-cost flow Metric Steiner tree preplanning Applying preplanning scheme to other problems: scheduling, multicommodity network flow, network design

36 Homework 1. (40 pts) In the paper, the authors argue that we can reduce the preplanning version of an NP-hard problem to solve a preplanning instance of another optimization problem that has a polynomial-time algorithm. Explain in your own words why this is true and give at least one example. 2. (a) (30 pts) Formulate the Minimum Steiner Tree problem as an optimization problem (b) (30 pts) Reformulate the Minimum Steiner Tree problem in a PCO (Preplanning Combinatorial Optimization) problem version. Explain the difference between (a) and (b).

37 Thank you … Any Questions?