Sketching complexity of graph cuts Alexandr Andoni joint work with: Robi Krauthgamer, David Woodruff.

Slides:



Advertisements
Similar presentations
1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Advertisements

Numerical Linear Algebra in the Streaming Model Ken Clarkson - IBM David Woodruff - IBM.
The Average Case Complexity of Counting Distinct Elements David Woodruff IBM Almaden.
Xiaoming Sun Tsinghua University David Woodruff MIT
Truthful Mechanisms for Combinatorial Auctions with Subadditive Bidders Speaker: Shahar Dobzinski Based on joint works with Noam Nisan & Michael Schapira.
Circuit and Communication Complexity. Karchmer – Wigderson Games Given The communication game G f : Alice getss.t. f(x)=1 Bob getss.t. f(y)=0 Goal: Find.
Exact Inference in Bayes Nets
A polylogarithmic approximation of the minimum bisection Robert Krauthgamer The Hebrew University Joint work with Uri Feige.
Approximating Average Parameters of Graphs Oded Goldreich, Weizmann Institute Dana Ron, Tel Aviv University.
All Rights Reserved © Alcatel-Lucent 2006, ##### Matthew Andrews, Alcatel-Lucent Bell Labs Princeton Approximation Workshop June 15, 2011 Edge-Disjoint.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo Department of Combinatorics and Optimization Joint work with Isaac.
Graph Sparsifiers: A Survey Nick Harvey Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman,
Graph Sparsifiers: A Survey Nick Harvey UBC Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman,
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey University of Waterloo Department of Combinatorics and Optimization Joint.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo C&O Joint work with Isaac Fung TexPoint fonts used in EMF. Read.
On Sketching Quadratic Forms Robert Krauthgamer, Weizmann Institute of Science Joint with: Alex Andoni, Jiecao Chen, Bo Qin, David Woodruff and Qin Zhang.
Sparsest Cut S S  G) = min |E(S, S)| |S| S µ V G = (V, E) c- balanced separator  G) = min |E(S, S)| |S| S µ V c |S| ¸ c ¢ |V| Both NP-hard.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 8 May 4, 2005
1 University of Freiburg Computer Networks and Telematics Prof. Christian Schindelhauer Distributed Coloring in Õ(  log n) Bit Rounds COST 293 GRAAL and.
Undirected ST-Connectivity 2 DL Omer Reingold, STOC 2005: Presented by: Fenghui Zhang CPSC 637 – paper presentation.
1 University of Freiburg Computer Networks and Telematics Prof. Christian Schindelhauer Mobile Ad Hoc Networks Theory of Data Flow and Random Placement.
Dept. of Computer Science Distributed Computing Group Asymptotically Optimal Mobile Ad-Hoc Routing Fabian Kuhn Roger Wattenhofer Aaron Zollinger.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint work with Mira Gonen Dana Ron Tel-Aviv University.
Distributed Combinatorial Optimization
Randomness in Computation and Communication Part 1: Randomized algorithms Lap Chi Lau CSE CUHK.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
How Robust are Linear Sketches to Adaptive Inputs? Moritz Hardt, David P. Woodruff IBM Research Almaden.
Approximating the MST Weight in Sublinear Time Bernard Chazelle (Princeton) Ronitt Rubinfeld (NEC) Luca Trevisan (U.C. Berkeley)
ICALP'05Stochastic Steiner without a Root1 Stochastic Steiner Trees without a Root Martin Pál Joint work with Anupam Gupta.
Approximation Algorithms for Stochastic Combinatorial Optimization Part I: Multistage problems Anupam Gupta Carnegie Mellon University.
Primal-Dual Meets Local Search: Approximating MST’s with Non-uniform Degree Bounds Author: Jochen Könemann R. Ravi From CMU CS 3150 Presentation by Dan.
Graph Sparsifiers Nick Harvey University of British Columbia Based on joint work with Isaac Fung, and independent work of Ramesh Hariharan & Debmalya Panigrahi.
1 Min-cut for Undirected Graphs Given an undirected graph, a global min-cut is a cut (S,V-S) minimizing the number of crossing edges, where a crossing.
Vertex Sparsification of Cuts, Flows, and Distances Robert Krauthgamer, Weizmann Institute of Science WorKer 2015, Nordfjordeid TexPoint fonts used in.
Tight Bounds for Graph Problems in Insertion Streams Xiaoming Sun and David P. Woodruff Chinese Academy of Sciences and IBM Research-Almaden.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Streaming Algorithms Piotr Indyk MIT. Data Streams A data stream is a sequence of data that is too large to be stored in available memory Examples: –Network.
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Spanning and Sparsifying Rajmohan Rajaraman Northeastern University, Boston May 2012 Chennai Network Optimization WorkshopSpanning and Sparsifying1.
Spectrally Thin Trees Nick Harvey University of British Columbia Joint work with Neil Olver (MIT  Vrije Universiteit) TexPoint fonts used in EMF. Read.
Sublinear Algorithms via Precision Sampling Alexandr Andoni (Microsoft Research) joint work with: Robert Krauthgamer (Weizmann Inst.) Krzysztof Onak (CMU)
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Artur Czumaj DIMAP DIMAP (Centre for Discrete Maths and it Applications) Computer Science & Department of Computer Science University of Warwick Testing.
Julia Chuzhoy (TTI-C) Yury Makarychev (TTI-C) Aravindan Vijayaraghavan (Princeton) Yuan Zhou (CMU)
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
Graph Partitioning using Single Commodity Flows
Lower bounds on data stream computations Seminar in Communication Complexity By Michael Umansky Instructor: Ronitt Rubinfeld.
Sampling in Graphs Alexandr Andoni (Microsoft Research)
The Message Passing Communication Model David Woodruff IBM Almaden.
Information Complexity Lower Bounds
Stochastic Streams: Sample Complexity vs. Space Complexity
New Characterizations in Turnstile Streams with Applications
Hans Bodlaender, Marek Cygan and Stefan Kratsch
Approximating the MST Weight in Sublinear Time
Lecture 18: Uniformity Testing Monotonicity Testing
Sublinear Algorithmic Tools 3
Approximating k-route cuts
Density Independent Algorithms for Sparsifying
MST in Log-Star Rounds of Congested Clique
Lecture 4: CountSketch High Frequencies
Lecture 7: Dynamic sampling Dimension Reduction
CIS 700: “algorithms for Big Data”
Instructor: Shengyu Zhang
The Communication Complexity of Distributed Set-Joins
CSCI B609: “Foundations of Data Science”
Sampling in Graphs: node sparsifiers
Lecture 6: Counting triangles Dynamic graphs & sampling
Dynamic Graph Algorithms
Complexity Theory: Foundations
Presentation transcript:

Sketching complexity of graph cuts Alexandr Andoni joint work with: Robi Krauthgamer, David Woodruff

Graph compression Why smaller graphs? faster algorithms use less storage space easier visualization

Preserve some structure Cuts approximately Other properties: Distances, (multi-commodity) flows, electrical conductance…

Problem definition Given G with n nodes and m edges For any set S: cut G (S)≈cut H (S) up to 1+ε approximation G H

Cut sparsifiers [Benczur-Karger’96]: can construct graph H with O(n/ε 2  log n) edges Approach: sample edges in the graph non-uniform sampling

Cut sparsifiers: constructions Ways to sample edges: strong connectivity [BK’96] connectivity, Nagamochi-Ibaraki index [FHHP’11] random spanning trees [GRV’09, FHHP’11] effective resistances [SS’08] random spanners [KP’12] Used to speed up cut problems [BK02, KL02, She09, Mad10] All: sparsifiers of size O(n/ε 2  polylog n) edges Can we do better? Yes: O(n/ε 2 ) edges [BSS’09]

Dependence on ε ? Factor of 1/ε can be expensive! Seems hard to avoid 1/ε 2 : for spectral sparsification: O(n/ε 2 ) is optimal [Alon-Boppana, Batson-Spielman-Srivastava’09] But still… for cut sparsifiers? above lower bounds are for complete graph we know how to estimate cuts in complete graphs!

Our results Lower bound: Ω(n/ε 2 ) bits required to represent a cut sparsifier Upper bound: O(n/ε  polylog n) sufficient for a relaxed notion! Relaxations: 1) instead of graph H => sketch H (data structure) 2) each “query” cut S preserved with high probability in contrast: all cuts at the same time G H S cut(S)

Motivating example

Upper Bound Three parts: i) sketch description ii) sketch size is O(n/ε  log 2 n) iii) estimation algorithm, correctness will focus on unweighted graphs

i) Sketch 1) guess coarse cut value V down-sample edges with probability 1/ε 2  1/V i.e., cuts of value ≈V become of value ≈1/ε 2 2) decompose along sparse cuts: in a connected component C, if exists P, |P|≤|C|/2, s.t. cut(P) ≤ 1/ε  |P| store cross-cut edges, (P, C\P) delete cross-cut edges from the graph, repeat 3) from remaining graph, store: connected (dense) components degree d x for each node x a sample of s=1/ε edges out of each node for each i=1,2,3… build sketch for guess V=2 i for each i=1,2,3… build sketch for guess V=2 i

ii) Sketch size Storing: Edges across sparse cuts Node degrees, connected components Sample of 1/ε edges out of each node Claim: total # edges across sparse cuts is O(n/ε · log n) sparse cut P => store ≤|P|·1/ε edges “charge” edges to nodes in P => 1/ε edges per node if node is charged, the size of its connected component halfs => can be charged only O(log n) times! Overall space: O(n/ε ·log n) · O(log n) O(n) space O(n/ ε ) space O(n/ ε log n ) space

iii) Estimation Given set S, need to estimate cut(S) Assume V=factor 2 guess of cut(S) estimate(S) = normalization * sum of # of sparse-cut edges in cut(S) for each dense component C: [sum of degrees inside S C ] – [estimate of # edges inside S C ] where S C = S intersection C (instead of: estimate of # of cross edges!) =ε 2 V (from down-sampling) =ε 2 V (from down-sampling)

Estimation illustration estimate(S) = normalization * sum of # of sparse-cut edges in cut(S) for each dense component C dense components C S SCSC

iii) Correctness Each of 3 steps preserves cut value up to 1+ε 1) down-sample to ≈1/ε 2 cut value approximation error 1+ε 2) edges across sparse cuts: no further approx! 3) edges inside a dense component? estimate of # edges inside S C has less variance than estimate of # edges across cut(S C )

Why not cross-cut edges?

Inside-edges estimate Estimate for component C: Claim 1: S C is small: |S C |≤2/ε cut(S C ) ≥ 1/ε ·|S C | (S C not a sparse cut) cut(S C ) ≤ 2/ε 2 (guessed cut value up to fac 2) hence |S C |≤2/ε SCSC

Inside edges estimate Claim 1: S C is small, |S C |≤2/ε Claim 2: Σ has standard deviation 3ε·cut(S C ) # edges inside S C can be large too, ≈O(1/ε 2 ) but cannot all be incident to the same node! #edges inside large only if |S C |≈1/ε nodes over all of them, 1/ε 2 sampled edges!

Wrap-up of upper bound Attentive variance calculation completes proof 1+O(ε) multiplicative approximation by Chebyshev inequality As usual: “high probability” by repeating logarithmic number of times Construction time? requires computing sparse cuts… NP-hard problem! OK to compute approximate sparse cuts! α-approximation => sketch size O(α · n/ε · log 2 n) E.g., using [Madry’10]: O(mn 1+δ /ε) runtime for O(n/ε · polylog n) space

An application Sketching for global min-cut Theorem: can compute sketch(G), of size O(n/ε · polylog n) so that: sketch(G 1 ), sketch(G 2 ) enough to compute global min-cut of G 1 +G 2 Obviously extends to >>2 partitions of graph Previous approaches have 1/ε 2 dependence [AG09, KL13, AGM12, GKK10, GKP12, KLMMS14]

Sketching for global min-cut Idea: store two “cut sparsifiers”: 1) relaxed: O(n/ε · polylog n) space 2) classic, for 1.5-approximation: O(n· log n) space Why enough? [Karger’94]: only ≤n 3 cuts within 1.5 of global min-cut list ≤n 3 such cuts using (2) for each, estimate up to 1+ε using (1) In general: sketch good for polynomial # of (non-adaptive) queries

Lower Bound Are relaxations necessary? graph H => sketch H each cut S preserved with high probability Theorem: if sketch H can 1+ε approximate the value of every cut S, then H must use Ω(n/ε 2 ) bits. Proof: via communication complexity need to use the fact that we can ask H an exponential number of cut queries!

Proof outline Assume: need to prove Ω(n 2 ) bits required Will solve Index(Gap-Ham) problem using sketch H then, Index(Gap-Ham) lower bound => H lower bound s 1, s 2 …s n : strings in {0,1} n t : string in {0,1} n index 1≤k≤n communication Problem: distinguish Δ(t,s k )<n/2-r(close) Δ(t,s k )>n/2+r(far) Problem: distinguish Δ(t,s k )<n/2-r(close) Δ(t,s k )>n/2+r(far)

Index(Gap-Ham) Essentially n-fold of Gap-Hamming Distribution: all s 1, s 2, …s n, t are random, conditioned on Δ(t,s i ) satisfying one of the 2 conditions Lemma: Index(Gap-Ham) requires Ω(n 2 ) communication s 1, s 2 …s n : strings in {0,1} n t : string in {0,1} n index 1≤k≤n communication Problem: distinguish Δ(t,s k )<n/2-r(close) Δ(t,s k )>n/2+r(far) Problem: distinguish Δ(t,s k )<n/2-r(close) Δ(t,s k )>n/2+r(far)

Index(Gap-Ham) via cut-sketch H Alice constructs a graph G (with 2n nodes) sends the cut-sketch H=sketch(G) to Bob Bob uses H to decide on Index(Gap-Ham) n=4 s1s1 s2s2 s3s3 s4s4 H string t => set T compute set S, of size n/2, minimizing cut(S U T) t=0011 if k inside S, output Δ(t,s k )<n/2-r

Intuition Hence, Bob chooses S that maximizes where Need to prove: k in S iff Δ(t,s k )<n/2-r n=4 s1s1 s2s2 s3s3 s4s4 H string t => set T compute set S, of size n/2, minimizing cut(S U T) if k inside S, output Δ(t,s k )<n/2-r

Intuition 2

Finale Cut sparsifiers of O(n/ε log 2 n) size under relaxations: 1) sketch (instead of a graph) 2) “with high probability” per query (instead of “for all”) relaxation (2) is necessary otherwise, need O(n/ε 2 ) size Open questions: 1)first relaxation necessary? i.e., estimate “#edges across cut” vs “sum degrees - #edges inside” 2)applications where “whp” guarantee enough ? polynomial # of (non-adaptive) queries? 3)same guarantee for spectral sparsification ? 1/ε 1.6 achievable [Chen, Qin, Woodruff, Zhang]