Metric Methods and Approximation Algorithms

Slides:



Advertisements
Similar presentations
Lower Bounds for Additive Spanners, Emulators, and More David P. Woodruff MIT and Tsinghua University To appear in FOCS, 2006.
Advertisements

Weighted Matching-Algorithms, Hamiltonian Cycles and TSP
Vertex sparsifiers: New results from old techniques (and some open questions) Robert Krauthgamer (Weizmann Institute) Joint work with Matthias Englert,
Great Theoretical Ideas in Computer Science
Lower Bound for Sparse Euclidean Spanners Presented by- Deepak Kumar Gupta(Y6154), Nandan Kumar Dubey(Y6279), Vishal Agrawal(Y6541)
Great Theoretical Ideas in Computer Science for Some.
Metric Embeddings with Relaxed Guarantees Hubert Chan Joint work with Kedar Dhamdhere, Anupam Gupta, Jon Kleinberg, Aleksandrs Slivkins.
1 The TSP : Approximation and Hardness of Approximation All exact science is dominated by the idea of approximation. -- Bertrand Russell ( )
Embedding Metrics into Ultrametrics and Graphs into Spanning Trees with Constant Average Distortion Ittai Abraham, Yair Bartal, Ofer Neiman The Hebrew.
Combinatorial Algorithms
Interchanging distance and capacity in probabilistic mappings Uriel Feige Weizmann Institute.
Navigating Nets: Simple algorithms for proximity search Robert Krauthgamer (IBM Almaden) Joint work with James R. Lee (UC Berkeley)
Great Theoretical Ideas in Computer Science.
1 Discrete Structures & Algorithms Graphs and Trees: II EECE 320.
Approximation Algorithms: Combinatorial Approaches Lecture 13: March 2.
1 Optimization problems such as MAXSAT, MIN NODE COVER, MAX INDEPENDENT SET, MAX CLIQUE, MIN SET COVER, TSP, KNAPSACK, BINPACKING do not have a polynomial.
Network Design Adam Meyerson Carnegie-Mellon University.
Approximating metrics by tree metrics Kunal Talwar Microsoft Research Silicon Valley Joint work with Jittat Fakcharoenphol Kasetsart University Thailand.
Great Theoretical Ideas in Computer Science.
An Approximation Algorithm for Requirement cut on graphs Viswanath Nagarajan Joint work with R. Ravi.
Universal Approximations for TSP, Steiner Tree, and Set Cover Authors: Lujun Jia, Guolong Lin, Guevara Noubir, Rajmohan Rajaraman, Ravi Sundaram Presented.
A General Approach to Online Network Optimization Problems Seffi Naor Computer Science Dept. Technion Haifa, Israel Joint work: Noga Alon, Yossi Azar,
Algorithms on negatively curved spaces James R. Lee University of Washington Robert Krauthgamer IBM Research (Almaden) TexPoint fonts used in EMF. Read.
Approximation Algorithms for Stochastic Combinatorial Optimization Part I: Multistage problems Anupam Gupta Carnegie Mellon University.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
1 The TSP : NP-Completeness Approximation and Hardness of Approximation All exact science is dominated by the idea of approximation. -- Bertrand Russell.
1 1 1-to-Many Distribution Vehicle Routing John H. Vande Vate Spring, 2005.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Randomized Online Algorithm for Minimum Metric Bipartite Matching Adam Meyerson UCLA.
Great Theoretical Ideas in Computer Science.
1 Oblivious Routing in Wireless networks Costas Busch Rensselaer Polytechnic Institute Joint work with: Malik Magdon-Ismail and Jing Xi.
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
Embeddings, flow, and cuts: an introduction University of Washington James R. Lee.
Universal Approximations in Network Design Rajmohan Rajaraman Northeastern University Based on joint work with Chinmoy Dutta, Lujun Jia, Guolong Lin, Jaikumar.
Doubling Dimension: a short survey Anupam Gupta Carnegie Mellon University Barriers in Computational Complexity II, CCI, Princeton.
Analysis of the Traveling Salesman Problem and current approaches for solving it. Rishi B. Jethwa and Mayank Agarwal. CSE Department. University of Texas.
Advances in Metric Embedding Theory Yair Bartal Hebrew University &Caltech UCLA IPAM 07.
Approximation Algorithms by bounding the OPT Instructor Neelima Gupta
1 Approximation Algorithms for Low- Distortion Embeddings into Low- Dimensional Spaces Badoiu et al. (SODA 2005) Presented by: Ethan Phelps-Goodman Atri.
Approximation algorithms
Hamiltonian Graphs Graphs Hubert Chan (Chapter 9.5)
Approximating Graphs by Trees Marcin Bieńkowski University of Wrocław
Information Complexity Lower Bounds
Optimization problems such as
Generalized Sparsest Cut and Embeddings of Negative-Type Metrics
Efficient methods for finding low-stretch spanning trees
Approximating the MST Weight in Sublinear Time
Hamiltonian Graphs Graphs Hubert Chan (Chapter 9.5)
Haim Kaplan and Uri Zwick
Polynomial-time approximation schemes for NP-hard geometric problems
Ultra-low-dimensional embeddings of doubling metrics
Great Theoretical Ideas in Computer Science
CS4234 Optimiz(s)ation Algorithms
Approximation Algorithms for TSP
Discrete Mathematics for Computer Science
Great Theoretical Ideas In Computer Science
Enumerating Distances Using Spanners of Bounded Degree
Lecture 16: Earth-Mover Distance
Near-Optimal (Euclidean) Metric Compression
Light Spanners for Snowflake Metrics
Introduction Wireless Ad-Hoc Network
Optimization Problems Online with Random Demands
Dimension versus Distortion a.k.a. Euclidean Dimension Reduction
Embedding Metrics into Geometric Spaces
Lecture 15: Least Square Regression Metric Embeddings
Advanced Algorithms Analysis and Design
Randomized Online Algorithm for Minimum Metric Bipartite Matching
The Intrinsic Dimension of Metric Spaces
Clustering.
Routing in Networks with Low Doubling Dimension
Presentation transcript:

Metric Methods and Approximation Algorithms Anupam Gupta Carnegie Mellon University 3ème cycle romand de Recherche Opérationnelle, 2010 Seminar, Lecture #1

y z x Metric space M = (V, d) set V of points symmetric non-negative distances d(x,y) triangle inequality d(x,y) ≤ d(x,z) + d(z,y) d(x,x) = 0 y z x

why metric spaces? Metric spaces are inputs to problems TSP round trip delays between machines distances between strings but also, Metric spaces are useful abstractions for various problems and interesting mathematical objects in their own right

Metrics 0 1 4 5 6 1 0 3 4 5 4 3 0 1 2 5 4 1 0 3 6 5 2 3 0 Distances d 5 e 3 d 6 a 5 2 1 1 3 4 c b 4

Choose your representation G F E D C B A - 468 44 126 454 85 69 347 170 514 509 137 527 450 127 348 90 493 46 114 387 208 497 58 188 388 282 513 436 115 331 151 404 244 327 94 234

Choose your representation SJ SD SF Sac PS Napa Mntr LA H.C. - 468 44 126 454 85 69 347 170 San Jose 514 509 137 527 450 127 348 San Diego 90 493 46 114 387 208 San Francisco 497 58 188 388 282 Sacramento 513 436 115 331 Palm Springs 151 404 244 327 94 Monterey 234 Hearst Castle

Choose your representation Good approximation via 2-d Euclidean distances or even via a sparse skeleton of edges: a “spanner” 497 234 115 127

Representations of metrics Just a distance matrix Shortest-path metric of a (“simple”) graph Points in Rk with distances given by the ℓp norms low-dimensional geometric representations

Tree metric a b a b d c c

Tree metric A metric M = (V,d) is a tree metric if there exists a tree T = (V [ X, E) (with edge-lengths) such that d = dT |V × V

more generally Given a family G of graphs A metric M = (V,d) is a G-metric if there exists a graph H = (V [ X, E) in G (with edge-lengths) such that d = dH |V × V

for now, G is the set of trees A metric M = (V,d) is a tree metric if there exists a tree T = (V [ X, E) (with edge-lengths) such that d = dT |V × V

Is every metric a tree metric? Consider the metric generated by the 4-cycle:

… “close” to a tree metric? What is “closeness” between metric spaces?

… “close” to a tree metric? Our favorite notion of closeness: Let M = (V,d) Let M’ = (V’,d’) Let f: V  V’ stretch(f) = maxx, y in V d(f(x), f(y)) d(x, y) d(f(x), f(y)) d(x, y) contraction(f) = maxx, y in V distortion(f) = stretch(f) × contraction(f) M is “close” to M’ if there exists a “low”-distortion map f: M  M’

examples

examples

Properties of distortion Invariant under scaling: If M is close to M’ and M’ is close to M’’, then…

is every metric close to a tree metric? So, does every metric admit a low-distortion embedding into a tree metric?

what do we do now? Here’s a way out… dT(a,b) ≥ d(a,b) E[dT(a,b)] ¼ ¼ ¼ c Here’s a way out… ¼ ¼ ¼ ¼ dT(a,b) ≥ d(a,b) E[dT(a,b)]

“dominating trees” Given a metric M = (V,d) let T(M) = { trees T | dT ≥ d } distances in such a tree T “dominates” distances in d

random tree embeddings Given a metric M = (V, d), want a probability distribution D over trees in T(M) such that 1) dT(x,y) ≥ d(x,y) for all x,y, all trees T in support(D) 2) for all x, y in V ET ← D[ dT(x,y) ] ≤ α d(x,y)

why are these useful? Consider, say, the traveling salesman problem:

quick recap of goals Given a metric M =(V,d) find a distribution over trees such that d(x,y) ≤ dT(x,y) for all trees in ET[ dT(x,y) ] ≤ ® × d(x,y)

first results [Alon Karp Peleg West ‘94] [Bartal ’96, ‘98]

current world record [Fakcharoenphol Rao Talwar ’03]

best possible (log n) lower bound for square grid hypercube diamond graphs constant-degree expander graphs

diamond graphs?

Time for some proofs…

roadmap To show an embedding of metrics into random trees, we show how to randomly partition metrics “well” how these random partitions give us good tree embeddings

useful notation Given a metric M = (V, d) Diameter(S) for S µ V Ball B(x,r)

high level desiderata Given a metric M, want to partition it randomly into pieces of “small” diameter such that “nearby” vertices lie in different pieces only with “small” probability. “random metric decompositions”

consider the path metric Suppose the pieces have diameter ¢ then the natural way to do it: gives Pr[ x, y in different clusters ] =

and the grid… Again, if the pieces have diameter ¢ then one natural way to do it: gives Pr[ x, y in different clusters ] ≈

“padded” decompositions A metric (V,d) admits ¯-padded decompositions, if for every ¢, we can output a random partition V = V1 ] V2 ] … ] Vk each Vj has diameter ≤ ¢ Pr[ x and y in different clusters ] ≤ 2’. Pr[ ball B(x,½) split ] ≤

“padded” decompositions A metric (V,d) admits ¯-padded decompositions, if for every ¢, we can output a random partition V = V1 ] V2 ] … ] Vk each Vj has diameter ≤ ¢ Pr[ B(x,½) split ] ≤

(weaker) theorems Theorem 1. Every n-point metric admits an ¯ = O(log n)-padded decomposition Theorem 2. [Bartal ’96] Embedding into distribution over trees with ® = O(log n × log “aspect-ratio”)

(stronger) theorems Theorem 3. Every n-point metric admits an ¯(x, ¢)-padded decomposition with Theorem 4. [FRT ’03] Embedding into distribution over trees with ® = O(log n)

we’ll prove the weaker results Theorem 1. Every n-point metric admits an ¯ = O(log n)-padded decomposition Theorem 2. [Bartal ’96] Embedding into distribution over trees with ® = O(log n × log “aspect-ratio”)

high-level idea

decomposition algorithm

decomposition algorithm each Vj has diameter ≤ ¢ Pr[ B(x,½) split ] ≤

proof (1) Fact #1: Pr[ any of the Rv values is more than ¢ ] < 1/poly(n)

proof (2) Fact #2: for any x in V, Pr[ Ball(x, ½) is split ] ≤

proof (3)

this is what we showed… each Vj has diameter ≤ ¢ Pr[ B(x,½) split ] ≤

Now to show Theorem 1. Every n-point metric admits an ¯ = O(log n)-padded decomposition Theorem 2. Embedding into distribution over trees with ® = O(log n × log aspect-ratio) we assume min distance = 1 (by scaling), so aspect-ratio = diameter

high-level idea

tree-building

tree-building Distances in tree are at least d(x,y) E[ distance(x,y) in tree ] ≤ O(log n log diam) d(x,y)

proof (1) Fact #1: Distances dT(x,y) are at least d(x,y) Proof: Suppose d(x,y) is in (2t-1, 2t]. Then x,y must be separated by RandT(*, i) for some i ≥ t And hence x and y are in different trees that have an edge of length at least 2i ≥ 2t ≥ d(x,y) between them. hence dT(x,y) ≥ 2i ≥ d(x,y)

proof (2) Fact #2: In any tree returned by RandT(X,i) distance of any node x from the root is ≤ 2i+1 Proof: By the algorithm, the distance is at most 2i + 2i-1 + … ≤ 2i+1. Corollary: If x,y in same tree returned by RandT(X, i), then dT(x,y) ≤ 2. 2i+1

proof (3) Fact #3: Expected distance E[ dT(x,y) ] ≤ O(log n log diam) d(x,y) Proof: If x,y separated at level t, the distance dT(x,y) is at most 4.2t (by Fact #2). The probability that x,y separated at level t is O(log n) d(x,y)/2t. Hence the expected distance is t 4.2t × O(log n) d(x,y)/2t = O(log n) d(x,y) × recursion depth = O(log n) d(x,y) × (log diameter)

now we’ve proved the full result Theorem 1. Every n-point metric admits an ¯ = O(log n)-padded decomposition Theorem 2. [Bartal ’96] Embedding into distribution over trees with ® = O(log n × log diameter)

extensions gives embedding into Hierarchically well-Separated Trees

extensions embedding special classes of graphs

extensions embedding graph metrics into distributions over their sub-trees

embeddings into trees

used for these problems k-median Group Steiner tree min-sum clustering metric labeling minimum communication spanning tree vehicle routing problems metrical task systems and k-server buy-at-bulk network design oblivious network design oblivious routing demand-cut problem …

app: “universal” TSP Given a metric (V,d), you find single permutation ¼ adversary gives you subset S µ V you use order given by ¼ to visit cities in S. How close are you to the optimal tour on S? If adversary does not look at actual permutation ¼ when choosing S  O(log n) factor worse in expectation. What if adversary can look at ¼ and then choose S? can use variant of tree embeddings to get O(log2 n)

recap… Two general techniques: Padded decompositions. Embeddings into random trees. Coming up: Embeddings into geometric space Dimension reduction for geometric point sets Dimension of an abstract metric space

thank you!