Graph Sparsifiers Nick Harvey University of British Columbia Based on joint work with Isaac Fung, and independent work of Ramesh Hariharan & Debmalya Panigrahi.

Slides:



Advertisements
Similar presentations
The Primal-Dual Method: Steiner Forest TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A AA A A A AA A A.
Advertisements

Matroid Bases and Matrix Concentration
Sparse Approximations
C&O 355 Lecture 23 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A A A A A.
Solving Laplacian Systems: Some Contributions from Theoretical Computer Science Nick Harvey UBC Department of Computer Science.
C&O 355 Mathematical Programming Fall 2010 Lecture 22 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Degree-3 Treewidth Sparsifiers Chandra Chekuri Julia Chuzhoy Univ. of IllinoisTTI Chicago.
C&O 355 Mathematical Programming Fall 2010 Lecture 21 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
Matrix Concentration Nick Harvey University of British Columbia TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
A polylogarithmic approximation of the minimum bisection Robert Krauthgamer The Hebrew University Joint work with Uri Feige.
Optimization of Pearl’s Method of Conditioning and Greedy-Like Approximation Algorithm for the Vertex Feedback Set Problem Authors: Ann Becker and Dan.
A Randomized Linear-Time Algorithm to Find Minimum Spanning Trees David R. Karger David R. Karger Philip N. Klein Philip N. Klein Robert E. Tarjan.
CSL758 Instructors: Naveen Garg Kavitha Telikepalli Scribe: Manish Singh Vaibhav Rastogi February 7 & 11, 2008.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo Department of Combinatorics and Optimization Joint work with Isaac.
Graph Sparsifiers: A Survey Nick Harvey Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman,
Graph Sparsifiers: A Survey Nick Harvey UBC Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman,
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey University of Waterloo Department of Combinatorics and Optimization Joint.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo C&O Joint work with Isaac Fung TexPoint fonts used in EMF. Read.
Finding Small Balanced Separators Author: Uriel Feige Mohammad Mahdian Presented by: Yang Liu.
Karger’s Min-Cut Algorithm Amihood Amir Bar-Ilan University, 2009.
Randomized Algorithms and Randomized Rounding Lecture 21: April 13 G n 2 leaves
Approximation Algorithm: Iterative Rounding Lecture 15: March 9.
Implicit Hitting Set Problems Richard M. Karp Harvard University August 29, 2011.
Greedy Algorithms Reading Material: Chapter 8 (Except Section 8.5)
An Approximation Algorithm for Requirement cut on graphs Viswanath Nagarajan Joint work with R. Ravi.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint work with Mira Gonen Dana Ron Tel-Aviv University.
Prize Collecting Cuts Daniel Golovin Carnegie Mellon University Lamps of ALADDIN 2005 Joint work with Mohit Singh & Viswanath Nagarajan.
Greedy Algorithms Like dynamic programming algorithms, greedy algorithms are usually designed to solve optimization problems Unlike dynamic programming.
Randomness in Computation and Communication Part 1: Randomized algorithms Lap Chi Lau CSE CUHK.
1 CSE 417: Algorithms and Computational Complexity Winter 2001 Lecture 11 Instructor: Paul Beame.
Approximation Algorithms: Bristol Summer School 2008 Seffi Naor Computer Science Dept. Technion Haifa, Israel TexPoint fonts used in EMF. Read the TexPoint.
C&O 355 Lecture 2 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A.
Approximating the MST Weight in Sublinear Time Bernard Chazelle (Princeton) Ronitt Rubinfeld (NEC) Luca Trevisan (U.C. Berkeley)
Minimal Spanning Trees What is a minimal spanning tree (MST) and how to find one.
Theory of Computing Lecture 10 MAS 714 Hartmut Klauck.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Simple and Improved Parameterized Algorithms for Multiterminal Cuts Mingyu Xiao The Chinese University of Hong Kong Hong Kong SAR, CHINA CSR 2008 Presentation,
The Best Algorithms are Randomized Algorithms N. Harvey C&O Dept TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A AAAA.
C&O 750 Randomized Algorithms Winter 2011 Lecture 24 Nicholas Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Approximating Minimum Bounded Degree Spanning Tree (MBDST) Mohit Singh and Lap Chi Lau “Approximating Minimum Bounded DegreeApproximating Minimum Bounded.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
An Algorithmic Proof of the Lopsided Lovasz Local Lemma Nick Harvey University of British Columbia Jan Vondrak IBM Almaden TexPoint fonts used in EMF.
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Spectrally Thin Trees Nick Harvey University of British Columbia Joint work with Neil Olver (MIT  Vrije Universiteit) TexPoint fonts used in EMF. Read.
Testing the independence number of hypergraphs
C&O 355 Mathematical Programming Fall 2010 Lecture 16 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Artur Czumaj DIMAP DIMAP (Centre for Discrete Maths and it Applications) Computer Science & Department of Computer Science University of Warwick Testing.
Graph Partitioning using Single Commodity Flows
Estimating PageRank on Graph Streams Atish Das Sarma (Georgia Tech) Sreenivas Gollapudi, Rina Panigrahy (Microsoft Research)
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
A randomized linear time algorithm for graph spanners Surender Baswana Postdoctoral Researcher Max Planck Institute for Computer Science Saarbruecken,
Correlation Clustering Nikhil Bansal Joint Work with Avrim Blum and Shuchi Chawla.
C&O 355 Lecture 19 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A A A A A.
Sketching complexity of graph cuts Alexandr Andoni joint work with: Robi Krauthgamer, David Woodruff.
Generating Random Spanning Trees via Fast Matrix Multiplication Keyulu Xu University of British Columbia Joint work with Nick Harvey TexPoint fonts used.
The Best Algorithms are Randomized Algorithms
Randomized Min-Cut Algorithm
Introduction to Algorithms
Approximating the MST Weight in Sublinear Time
From dense to sparse and back again: On testing graph properties (and some properties of Oded)
Density Independent Algorithms for Sparsifying
MST in Log-Star Rounds of Congested Clique
CIS 700: “algorithms for Big Data”
Randomized Algorithms CS648
Matrix Martingales in Randomized Numerical Linear Algebra
Chapter 23 Minimum Spanning Tree
CSCI B609: “Foundations of Data Science”
Lecture 14 Shortest Path (cont’d) Minimum Spanning Tree
Lecture 13 Shortest Path (cont’d) Minimum Spanning Tree
Presentation transcript:

Graph Sparsifiers Nick Harvey University of British Columbia Based on joint work with Isaac Fung, and independent work of Ramesh Hariharan & Debmalya Panigrahi TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A

Approximating Dense Objects by Sparse Objects Floor joists Wood JoistsEngineered Joists

Approximating Dense Objects by Sparse Objects Bridges Masonry ArchTruss Arch

Approximating Dense Objects by Sparse Objects Bones Human FemurRobin Bone

Approximating Dense Objects by Sparse Objects Graphs Dense GraphSparse Graph

Input: Undirected graph G=(V,E), weights u : E ! R + Output: A subgraph H=(V,F) of G with weights w : F ! R + such that |F| is small and u( ± G (U)) = (1 § ² ) w( ± H (U)) 8 U µ V weight of edges between U and V\U in Gweight of edges between U and V\U in H Graph Sparsifiers Weighted subgraphs that approximately preserve graph structure Cut Sparsifiers (Karger ‘94) GH

Input: Undirected graph G=(V,E), weights u : E ! R + Output: A subgraph H=(V,F) of G with weights w : F ! R + such that |F| is small and x T L G x = (1 § ² ) x T L H x 8 x 2 R V Spectral Sparsifiers Weighted subgraphs that approximately preserve graph structure (Spielman-Teng ‘04) Laplacian matrix of GLaplacian matrix of H GH

Motivation: Faster Algorithms Dense Input graph G Exact/Approx Output Algorithm A for some problem P Sparse graph H Approximately preserves solution of P Algorithm A runs faster on sparse input Approximate Output (Fast) Sparsification Algorithm Min s-t cut, Sparsest cut, Max cut, …

State of the art SparsifierAuthors# edgesRunning Time CutBenczur-Karger ’96O(n log n / ² 2 )O(m log 3 n) SpectralSpielman-Teng ’04O(n log c n / ² 2 )O(m log c n / ² 2 ) SpectralSpielman-Srivastava ’08O(n log n / ² 2 )O(m log c n / ² 2 ) SpectralBatson et al. ’09O(n / ² 2 )O(mn 3 ) CutThis paperO(n log n / ² 2 )O(m + n log 5 n / ² 2 ) * SpectralLevin-Koutis-Peng ’12O(n log n / ² 2 )O(m log 2 n / ² 2 ) n = # vertices m = # edges *: The best algorithm in our paper is due to Panigrahi. ~ c = large constant

Random Sampling Can’t sample edges with same probability! Idea: [Benczur-Karger ’96] Sample low-connectivity edges with high probability, and high-connectivity edges with low probability Keep this Eliminate most of these

Generic algorithm Input: Graph G=(V,E), weights u : E ! R + Output: A subgraph H=(V,F) with weights w : F ! R + Choose ½ (= #sampling iterations) Choose probabilities { p e : e 2 E } For i=1 to ½ For each edge e 2 E With probability p e Add e to F Increase w e by u e /( ½ p e ) E[|F|] · ½ ¢  e p e E[ w e ] = u e 8 e 2 E ) For every U µ V, E[ w( ± H (U)) ] = u( ± G (U)) [Benczur-Karger ‘96] Goal 1: E[|F|] = O(n log n / ² 2 ) How should we choose these parameters? Goal 2: w( ± H (U)) is highly concentrated

Benczur-Karger Algorithm Input: Graph G=(V,E), weights u : E ! R + Output: A subgraph H=(V,F) with weights w : F ! R + Choose ½ = O(log n / ² 2 ) Let p e = 1/“strength” of edge e For i=1 to ½ For each edge e 2 E With probability p e Add e to F Increase w e by u e /( ½ p e ) Fact 1: E[|F|] = O(n log n / ² 2 ) Fact 2: w( ± H (U)) is very highly concentrated ) For every U µ V, w( ± H (U)) = (1 § ² ) u( ± G (U)) “strength” is a slightly unusual quantity, but Fact 3: Can estimate all edge strengths in O(m log 3 n) time “strength” is a slightly unusual quantity Question: [BK ‘02] Can we use connectivity instead of strength?

Our Algorithm Input: Graph G=(V,E), weights u : E ! R + Output: A subgraph H=(V,F) with weights w : F ! R + Choose ½ = O(log 2 n / ² 2 ) Let p e = 1/“connectivity” of e For i=1 to ½ For each edge e 2 E With probability p e Add e to F Increase w e by u e /( ½ p e ) Fact 1: E[|F|] = O(n log 2 n / ² 2 ) Fact 2: w( ± H (U)) is very highly concentrated ) For every U µ V, w( ± H (U)) = (1 § ² ) u( ± G (U)) Extra trick: Can shrink |F| to O(n log n / ² 2 ) by using Benczur-Karger to sparsify our sparsifier!

Motivation for our algorithm Connectivities are simpler and more natural ) Faster to compute Fact: Can estimate all edge connectivities in O(m + n log n) time [Ibaraki-Nagamochi ’92] ) Useful in other scenarios Our sampling method has been used to compute sparsifiers in the streaming model [Ahn-Guha-McGregor ’12]

Overview of Analysis Most cuts hit a huge number of edges ) extremely concentrated ) whp, most cuts are close to their mean Most Cuts are Big & Easy!

Overview of Analysis High connectivity Low sampling probability Low connectivity High sampling probability Hits many red edges ) reasonably concentrated Hits only one red edge ) poorly concentrated The same cut also hits many green edges ) highly concentrated This masks the poor concentration above There are few small cuts [Karger ’94], so probably all are concentrated. Key Question: Are there few such cuts? Key Lemma: Yes!

Notation: k uv = min size of a cut separating u and v Main ideas: – Partition edges into connectivity classes E = E 1 [ E 2 [... E log n where E i = { e : 2 i-1 · k e <2 i } – Prove weight of sampled edges that each cut takes from each connectivity class is about right – This yields a sparsifier U

Prove weight of sampled edges that each cut takes from each connectivity class is about right Notation: C = ± (U) is a cut C i := ± (U) Å E i is a cut-induced set Chernoff bounds can analyze each cut-induced set, but… Key Question: Are there few small cut-induced sets? C1C1 C2C2 C3C3 C4C4

Counting Small Cuts Lemma: [Karger ’93] Let G=(V,E) be a graph. Let K be the edge-connectivity of G. (i.e., global min cut value) Then, for every ® ¸ 1, |{ ± (U) : | ± (U)| · ® K }| < n 2 ®. Example: Let G = n-cycle. Edge connectivity is K=2. Number of cuts of size c = £ ( n c ). ) |{ ± (U) : | ± (U)| · ® K }| · O (n 2 ® ).

Counting Small Cut-Induced Sets Our Lemma: Let G=(V,E) be a graph. Fix any B µ E. Suppose k e ¸ K for all e in B. (k uv = min size of a cut separating u and v) Then, for every ® ¸ 1, |{ ± (U) Å B : | ± (U)| · ® K }| < n 2 ®. Karger’s Lemma: the special case B=E and K=min cut.

When is Karger’s Lemma Weak? Lemma: [Karger ’93] Let G=(V,E) be a graph. Let K be the edge-connectivity of G. (i.e., global min cut value) Then, for every c ¸ K, |{ ± (U) : | ± (U)| · c }| < n 2c/K. Example: Let G = n-cycle. Edge connectivity is K=2 |{ cuts of size c }| < n c ² K = ² < n 2c/ ²

Our Lemma Still Works Our Lemma: Let G=(V,E) be a graph. Fix any B µ E. Suppose k e ¸ K for all e in B. (k uv = min size of a cut separating u and v) Then, for every ® ¸ 1, |{ ± (U) Å B : | ± (U)| · ® K }| < n 2 ®. Example: Let G = n-cycle. Let B = cycle edges. We can take K=2. So |{ ± (U) Å B : | ± (U)| · ® K }| < n 2 ®. |{ cut-induced subsets of B induced by cuts of size · c }| · n c ²

Algorithm for Finding a Min Cut [Karger ’93] Input: A graph Output: A minimum cut (maybe) While graph has  2 vertices – Pick an edge at random – Contract it End While Output remaining edges Claim: For any min cut, this algorithm outputs it with probability ¸ 1/n 2. Corollary: There are · n 2 min cuts.

Finding a Small Cut-Induced Set Input: A graph G=(V,E), and B µ E Output: A cut-induced subset of B While graph has  2 vertices – If some vertex v has no incident edges in B Split-off all edges at v and delete v – Pick an edge at random – Contract it End While Output remaining edges in B Claim: For any min cut-induced subset of B, this algorithm outputs it with probability > 1/n 2. Corollary: There are < n 2 min cut-induced subsets of B Splitting Off Replace edges {u,v} and {u’,v} with {u,u’} while preserving edge-connectivity between all vertices other than v Splitting Off Replace edges {u,v} and {u’,v} with {u,u’} while preserving edge-connectivity between all vertices other than v Wolfgang Mader v u u’ v u

Conclusions Questions Sampling according to connectivities gives a sparsifier We generalize Karger’s cut counting lemma Improve O(log 2 n) to O(log n) in sampling analysis Applications of our cut-counting lemma?