Small Set Expansion in The Johnson Graph

Slides:



Advertisements
Similar presentations
Lower Bounds for Additive Spanners, Emulators, and More David P. Woodruff MIT and Tsinghua University To appear in FOCS, 2006.
Advertisements

Linear-Degree Extractors and the Inapproximability of Max Clique and Chromatic Number David Zuckerman University of Texas at Austin.
Shortest Vector In A Lattice is NP-Hard to approximate
Inapproximability of Hypergraph Vertex-Cover. A k-uniform hypergraph H= : V – a set of vertices E - a collection of k-element subsets of V Example: k=3.
Inapproximability of MAX-CUT Khot,Kindler,Mossel and O ’ Donnell Moshe Ben Nehemia June 05.
Gillat Kol joint work with Ran Raz Locally Testable Codes Analogues to the Unique Games Conjecture Do Not Exist.
Dana Moshkovitz MIT Joint work with Subhash Khot, NYU 1.
Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
MaxClique Inapproximability Seminar on HARDNESS OF APPROXIMATION PROBLEMS by Dr. Irit Dinur Presented by Rica Gonen.
Parallel Repetition From Fortification Dana Moshkovitz MIT.
Dana Moshkovitz. Back to NP L  NP iff members have short, efficiently checkable, certificates of membership. Is  satisfiable?  x 1 = truex 11 = true.
Introduction to PCP and Hardness of Approximation Dana Moshkovitz Princeton University and The Institute for Advanced Study 1.
The PCP Theorem via gap amplification Irit Dinur Hebrew University.
1/17 Optimal Long Test with One Free Bit Nikhil Bansal (IBM) Subhash Khot (NYU)
Gillat Kol joint work with Ran Raz Locally Testable Codes Analogues to the Unique Games Conjecture Do Not Exist.
Two Query PCP with Sub-constant Error Dana Moshkovitz Princeton University Ran Raz Weizmann Institute 1.
Random Walks Ben Hescott CS591a1 November 18, 2002.
Asaf Shapira (Georgia Tech) Joint work with: Arnab Bhattacharyya (MIT) Elena Grigorescu (Georgia Tech) Prasad Raghavendra (Georgia Tech) 1 Testing Odd-Cycle.
Derandomized DP  Thus far, the DP-test was over sets of size k  For instance, the Z-Test required three random sets: a set of size k, a set of size k-k’
1 Mazes In The Theory of Computer Science Dana Moshkovitz.
Michael Bender - SUNY Stony Brook Dana Ron - Tel Aviv University Testing Acyclicity of Directed Graphs in Sublinear Time.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
Linear-Time Encodable and Decodable Error-Correcting Codes Jed Liu 3 March 2003.
Dana Moshkovitz, MIT Joint work with Subhash Khot, NYU.
Wei Wang Xi’an Jiaotong University Generalized Spectral Characterization of Graphs: Revisited Shanghai Conference on Algebraic Combinatorics (SCAC), Shanghai,
Correlation testing for affine invariant properties on Shachar Lovett Institute for Advanced Study Joint with Hamed Hatami (McGill)
Sub-Constant Error Low Degree Test of Almost-Linear Size Dana Moshkovitz Weizmann Institute Ran Raz Weizmann Institute.
5.2 Trees  A tree is a connected graph without any cycles.
15-853:Algorithms in the Real World
Ramsey Properties of Random Graphs; A Sharp Threshold Proven via A Hypergraph Regularity Lemma. Ehud Friedgut, Vojtech Rödl, Andrzej Rucinski, Prasad.
1/19 Minimizing weighted completion time with precedence constraints Nikhil Bansal (IBM) Subhash Khot (NYU)
1 ۞ An eigenvalue λ and an eigenfunction f(x) of an operator Ĥ in a space S satisfy Week 6 2. Properties of self-adjoint operators where f(x) is implied.
Presented by Alon Levin
The Message Passing Communication Model David Woodruff IBM Almaden.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
Random Access Codes and a Hypercontractive Inequality for
From Classical Proof Theory to P vs. NP
Markov Chains and Random Walks
New Characterizations in Turnstile Streams with Applications
Standard Representations of Regular Languages
CSE322 PUMPING LEMMA FOR REGULAR SETS AND ITS APPLICATIONS
Markov Chains Mixing Times Lecture 5
PROPERTIES OF REGULAR LANGUAGES
Lecture 9 Theory of AUTOMATA
Approximate Matchings in Dynamic Graph Streams
Lecture 18: Uniformity Testing Monotonicity Testing
Dana Moshkovitz The Institute For Advanced Study
Log-Sobolev Inequality on the Multislice (and what those words mean)
Direct product testing
Structural Properties of Low Threshold Rank Graphs
Enumerating Distances Using Spanners of Bounded Degree
Subhash Khot Theory Group
Haim Kaplan and Uri Zwick
Depth Estimation via Sampling
Linear sketching with parities
The Effect of Network Topology on the Spread of Epidemics
Kleene’s Theorem Muhammad Arif 12/6/2018.
Linear sketching over
Introduction to PCP and Hardness of Approximation
On the effect of randomness on planted 3-coloring models
Ilan Ben-Bassat Omri Weinstein
CS21 Decidability and Tractability
Controllability and Observability of Linear Dynamical Equations
The Weizmann Institute
Kleene’s Theorem (Part-3)
Switching Lemmas and Proof Complexity
Sparse Kindler-Safra Theorem via agreement theorems
Integer and fractional packing of graph families
Small Set Expansion in The Johnson Graph
Presentation transcript:

Small Set Expansion in The Johnson Graph Subhash Khot (NYU), Dor Minzer (TAU), Dana Moshkovitz (UT Austin), Muli Safra (TAU)

Hard to Approximate Input The Johnson Graph Definition (Johnson graph): The nodes are sets of size K in a universe of size N (typically, K<<N). Two sets are connected if they intersect in L elements (typically, L=(K)). Exploring the Johnson graph was a crucial step towards the analysis of the related Grassmann graph and the recent proof of the 2-to-2 Conjecture Hard to Approximate Input

The Johnson Graph As Slice of Noisy Hypercube Noisy hypercube graph (variant): the vertices are strings in {0,1}N. There is an edge between x and y if they differ on N coordinates. Johnson graphs are “slices” of noisy hypercube: restrict to strings with K ones. Edges correspond to intersection of size L=(1-)K. N 1 Prove a “Fourier analytic” result for slices analogous to a fundamental result for hypercube

The Johnson Graph Underlies Direct Product Testing We obtain an optimal direct product testing theorem (if the verifier accepts with non-negligible probability, players must agree on N bits) Direct Product Test: Two players supposedly agree on N bits. A verifier picks at random two sets S,T[N], |S|=|T|=k, |ST|=L. The verifier sends S to one player, T to the other player. Each player is supposed to send the K bits indicated by their set. The verifier checks that the answers on L common bits are the same. S S T T

S Expansion Let G=(V,E) be a d-regular graph. The expansion of SV is (S) = |E(S,V-S)| / (d|S|). Small Set Expander: (S)  1- for all S of size < |V|. Examples: The noisy hypercube is a small set expander by hypercontractivity. The Johnson graph is not a small set expander (K<<N, L/K large). S

Johnson Graph - Not Small Set Expander Dictator The Zoom-in family: Pick an element x and consider the family S of all sets that contain x. |S|/|(Nk)|  (K/N) [small for K<<N] (S)  1 – (L/K) [much smaller than 1 for constant L/K]. Rough Statement of Main Theorem: Except for “dictator like families”, all small families S in the Johnson graph have (S)1-o(1).

|PAR(AS) - PA(AS) |   Unlike Dictator Dfn: A family S of sets is (r,)-pseudorandom if for all R[N], |R|r, |PAR(AS) - PA(AS) |   Examples: Dictator is not pseudorandom: for S = {A that contains x} take R={x}, so PAR(AS)=1, while PA(AS)K/N. A sufficiently large random family S is pseudorandom with high probability. (S)>1 – (L/K)r – rO(r)poly(1/) – poly(2K3,1/)/N1/16 Main Theorem: Let S be of fraction at most  and (r,)-pseudorandom. For sufficiently large N (as a function of K and 1/), (S)>1 – (L/K)r – 2O(r)poly() – o(1).

Approach J F The proof is via spectral analysis. 𝑁 𝐾 The proof is via spectral analysis. Let J be the transition matrix of the Johnson graph. Let F be the indicator of a small pseudorandom family. Want to show <F,JF> is small. J F 𝑁 𝐾

Level Picture K K-1 … r r-1 From Main Lem To Main Thm: The Johnson transition matrix J has K+1 eigenspaces; the i’th has eigenvalue roughly (L/K)i. Main Lemma: Small pseudorandom families F= ∑i Fi concentrated on upper levels, i.e., ∑ir|Fi|22<0.01|F|22. K K-1 … r r-1 From Main Lem To Main Thm: Let F be small pseudorandom. Write F=∑i Fi with Fi in i‘th space. <F,JF> = ∑i <Fi,JFi>  ∑i>r (L/K)i|Fi|22  (L/K)r|F|22.

The Eigenspaces A We say that a function F from sets of size K to the reals is of degree i if there exists f from sets of size i to the reals such that: F(A) = ∑IA,|I|=i f(I). For example: F(A) = “does A contain x?” is of degree 1: take f({y})=“does y=x?”. Lemma: Let Si be the space of functions of degree i that are orthogonal to the functions of degree i-1. Then Si is an eigenspace of J with eigenvalue  (L/K)i.

Level Picture K K-1 … r ... 1 Zoom-in/Dictators The Johnson transition matrix J has K+1 eigenspaces; the i’th corresponds to degree i; has eigenvalue roughly (L/K)i. Main Lemma: Small pseudorandom families F= ∑i Fi concentrated on upper levels, i.e., ∑ir|Fi|22<0.01|F|22. K K-1 … r ... 1 For simplicity, we’ll prove that F is not almost entirely in one level i, i.e., |Fi|22<0.99|F|22 Zoom-in/Dictators

Fourth Moment Lemma: Let N be sufficiently large Fourth Moment Lemma: Let N be sufficiently large. Let F be of fraction  and (r,)-pseudorandom with F=F0+…+FK where Fi is in the i‘th eigenspace. Then for i=0,…,r, we have EA[Fi(A)4]  2O(i)  EA[Fi(A)2] + o(1). F not almost entirely on one level 0ir: Assume on way of contradiction that |Fi|22 0.99|F|22=0.99. Then F(A)Fi(A) for all but at most 0.01 fraction of A, and since F is a set of fraction  we have EA[Fi(A)4] EA[F(A)4]=EA[F(A)2] EA[Fi(A)2] , which contradicts the Fourth Moment Lemma. (More generally, Fourth Moment Lemma implies EA[Fi(A)2]  2O(i) 1/4 + o(1) for i=0,…,r, which implies the main lemma) Fourth Moment Lemma: Let N be sufficiently large. Let F be (r,)-pseudorandom with F=F0+…+FK where Fi is in the i‘th eigenspace. Then for i=0,…,r, we have EA[Fi(A)4]  iO(i)  EA[Fi(A)2] + o(1). Claim: Let =EA[F(A)] and i=EA[Fi(A)2]. Then PA( Fi(A) ≥ i/(4) ) ≥ i/2. Claim  EA[Fi(A)4] ≥ i5/(294)  By 4th Moment Lem, EA[Fi(A)2]  iO(i) 1/4 + o(1), which implies main lemma. Proof of Claim: EA[F(A)2]=  By orthogonality, EA[(F(A)-Fi(A))2]=-i. By Markov, PA( (F(A)-Fi(A))2 ≥ 1-i/(2) )  (-i)/(1-i/(2))  -i/2.

The Main Technical Difficulty Fourth Moment Lemma: Let N be sufficiently large. Let F be of fraction  and (r,)-pseudorandom with F=F0+…+FK where Fi is in the i‘th eigenspace. Then for i=0,…,r, we have EA[Fi(A)4]  2O(i)  EA[Fi(A)2] + o(1). Write Fi(A)=∑IA,|I|=i f(I). EA[Fi(A)4]=EA[∑I1,I2,I3,I4A,|Ij|=i f(I1)f(I2)f(I3)f(I4)]. I2 How to bound correlations?? I1 I4 I3

(maxx2Ex3,x1[f({x1,x2,x3})2])1/2 Ex1,x2,x3[f({x1,x2,x3})2] For simplicity, assume i=3 and consider third moments. Fix the “intersection pattern” (How many elements in I1I2I3? How many in I1cI2I3? …). For simplicity, focus on the pattern below. Cauchy-Schwarz & Restrict technique: Can be shown to be proportional to: (maxx2Ex3,x1[f({x1,x2,x3})2]Ex2,x4,x3[f({x2,x3,x4})2]Ex2,x4,x1[f({x1,x2,x4})2])1/2 (maxx2Ex3,x1[f({x1,x2,x3})2])1/2 Ex1,x2,x3[f({x1,x2,x3})2] Ex2[Ex3,x1[f({x1,x2,x3})2]Ex4,x3[f({x2,x3,x4})2]]1/2Ex2,x4,x1[f({x1,x2,x4})2]1/2 Ex2[Ex3,x1[f({x1,x2,x3})2]1/2Ex4,x3[f({x2,x3,x4})2]1/2Ex4,x1[f({x1,x2,x4})2]1/2] Ex2[Ex3,x1[f({x1,x2,x3})2]1/2Ex4[Ex3[f({x2,x3,x4})2]1/2Ex1[f({x1,x2,x4})2]1/2]] Ex2,x4[Ex3,x1[f({x1,x2,x3})2]1/2Ex3[f({x2,x3,x4})2]1/2Ex1[f({x1,x2,x4})2]1/2] Ex2,x3,x4[Ex1[f({x1,x2,x3})2]1/2 Ex1[f({x1,x2,x4})2]1/2f({x2,x3,x4})] Ex2,x4[Ex3[Ex1[f({x1,x2,x3})2]1/2 f({x2,x3,x4})]Ex1[f({x1,x2,x4})2]1/2] Ex2,x3,x4[Ex1[f({x1,x2,x3})f({x1,x2,x4})]f({x2,x3,x4})] Ex1,x2,x3,x4[f({x1,x2,x3})f({x1,x2,x4})f({x2,x3,x4})] I2 (maxxEB[Fi(B{x})2])1/2 EA[Fi(A)2] I1 Upper bounded by O() from pseudorandomness I3