Dana Moshkovitz MIT Joint work with Subhash Khot, NYU 1.

Slides:



Advertisements
Similar presentations
1/15 Agnostically learning halfspaces FOCS /15 Set X, F class of functions f: X! {0,1}. Efficient Agnostic Learner w.h.p. h: X! {0,1} poly(1/ )
Advertisements

3.6 Support Vector Machines
Constraint Satisfaction Problems
Hardness of Reconstructing Multivariate Polynomials. Parikshit Gopalan U. Washington Parikshit Gopalan U. Washington Subhash Khot NYU/Gatech Rishi Saket.
Subspace Embeddings for the L1 norm with Applications Christian Sohler David Woodruff TU Dortmund IBM Almaden.
Analysis of Algorithms
Introduction to Algorithms 6.046J/18.401J/SMA5503
Tuesday, March 5 Duality – The art of obtaining bounds – weak and strong duality Handouts: Lecture Notes.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
FACTORING ax2 + bx + c Think “unfoil” Work down, Show all steps.
1 Approximability Results for Induced Matchings in Graphs David Manlove University of Glasgow Joint work with Billy Duckworth Michele Zito Macquarie University.
Uri Feige Microsoft Understanding Parallel Repetition Requires Understanding Foams Guy Kindler Weizmann Ryan ODonnell CMU.
1 LP, extended maxflow, TRW OR: How to understand Vladimirs most recent work Ramin Zabih Cornell University.
Reductions Complexity ©D.Moshkovitz.
1 Outline relationship among topics secrets LP with upper bounds by Simplex method basic feasible solution (BFS) by Simplex method for bounded variables.
Comp 122, Spring 2004 Order Statistics. order - 2 Lin / Devi Comp 122 Order Statistic i th order statistic: i th smallest element of a set of n elements.
Mathematics for Economics Beatrice Venturi 1 Economics Faculty CONTINUOUS TIME: LINEAR DIFFERENTIAL EQUATIONS Economic Applications LESSON 2 prof. Beatrice.
1 The tiling algorithm Learning in feedforward layered networks: the tiling algorithm writed by Marc M é zard and Jean-Pierre Nadal.
1 Column Generation. 2 Outline trim loss problem different formulations column generation the trim loss problem master problem and subproblem in column.
Randomized Algorithms Randomized Algorithms CS648 1.
Recurrences : 1 Chapter 3. Growth of function Chapter 4. Recurrences.
5-1 Chapter 5 Theory & Problems of Probability & Statistics Murray R. Spiegel Sampling Theory.
MAT 205 F08 Chapter 12 Complex Numbers.
Exploring the border between P and NP Uriel Feige Weizmann Institute 1.
How to Round Any CSP Prasad Raghavendra University of Washington, Seattle David Steurer, Princeton University (In Principle)
Weighted moving average charts for detecting small shifts in process mean or trends The wonders of JMP 1.
Hypothesis Tests: Two Independent Samples
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Jan, 29, 2014.
25 seconds left…...
Linear Round Integrality Gaps for the Lasserre Hierarchy Grant Schoenebeck.
Complexity ©D.Moshkovits 1 Where Can We Draw The Line? On the Hardness of Satisfiability Problems.
Shortest Vector In A Lattice is NP-Hard to approximate
Subhash Khot IAS Elchanan Mossel UC Berkeley Guy Kindler DIMACS Ryan O’Donnell IAS.
Statistical Inferences Based on Two Samples
Abdollah Khodkar Department of Mathematics University of West Georgia Joint work with Arezoo N. Ghameshlou, University of Tehran.
Chapter 8 Estimation Understandable Statistics Ninth Edition
Approximate List- Decoding and Hardness Amplification Valentine Kabanets (SFU) joint work with Russell Impagliazzo and Ragesh Jaiswal (UCSD)
PSSA Preparation.
9. Two Functions of Two Random Variables
From Approximative Kernelization to High Fidelity Reductions joint with Michael Fellows Ariel Kulik Frances Rosamond Technion Charles Darwin Univ. Hadas.
1 Complexity ©D.Moshkovitz Cryptography Where Complexity Finally Comes In Handy…
Bart Jansen 1.  Problem definition  Instance: Connected graph G, positive integer k  Question: Is there a spanning tree for G with at least k leaves?
1 Decidability continued…. 2 Theorem: For a recursively enumerable language it is undecidable to determine whether is finite Proof: We will reduce the.
Inapproximability of MAX-CUT Khot,Kindler,Mossel and O ’ Donnell Moshe Ben Nehemia June 05.
Approximability & Sums of Squares Ryan O’Donnell Carnegie Mellon.
Chapter 5 The Mathematics of Diversification
all-pairs shortest paths in undirected graphs
Minimum Vertex Cover in Rectangle Graphs
Discrepancy and SDPs Nikhil Bansal (TU Eindhoven, Netherlands ) August 24, ISMP 2012, Berlin.
Gillat Kol joint work with Ran Raz Locally Testable Codes Analogues to the Unique Games Conjecture Do Not Exist.
The Max-Cut problem: Election recounts? Majority vs. Electoral College? 7812.
On the Unique Games Conjecture Subhash Khot Georgia Inst. Of Technology. At FOCS 2005.
Inapproximability Seminar – 2005 David Arnon  March 3, 2005 Some Optimal Inapproximability Results Johan Håstad Royal Institute of Technology, Sweden.
MaxClique Inapproximability Seminar on HARDNESS OF APPROXIMATION PROBLEMS by Dr. Irit Dinur Presented by Rica Gonen.
Constraint Satisfaction over a Non-Boolean Domain Approximation Algorithms and Unique Games Hardness Venkatesan Guruswami Prasad Raghavendra University.
Introduction to PCP and Hardness of Approximation Dana Moshkovitz Princeton University and The Institute for Advanced Study 1.
A 3-Query PCP over integers a.k.a Solving Sparse Linear Systems Prasad Raghavendra Venkatesan Guruswami.
1/17 Optimal Long Test with One Free Bit Nikhil Bansal (IBM) Subhash Khot (NYU)
Inapproximability from different hardness assumptions Prahladh Harsha TIFR 2011 School on Approximability.
Dictator tests and Hardness of approximating Max-Cut-Gain Ryan O’Donnell Carnegie Mellon (includes joint work with Subhash Khot of Georgia Tech)
Venkatesan Guruswami (CMU) Yuan Zhou (CMU). Satisfiable CSPs Theorem [Schaefer'78] Only three nontrivial Boolean CSPs for which satisfiability is poly-time.
Dana Moshkovitz, MIT Joint work with Subhash Khot, NYU.
1/19 Minimizing weighted completion time with precedence constraints Nikhil Bansal (IBM) Subhash Khot (NYU)
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
Sum of Squares, Planted Clique, and Pseudo-Calibration
Possibilities and Limitations in Computation
Subhash Khot Dept of Computer Science NYU-Courant & Georgia Tech
Introduction to PCP and Hardness of Approximation
Small Set Expansion in The Johnson Graph
Presentation transcript:

Dana Moshkovitz MIT Joint work with Subhash Khot, NYU 1

Best Approximations Known For 2CSP vs. 3CSP 23 OR [FG95] 7/8 XOR [GW92] 1/2 AND [FG95] 1/2 any better approximation: NP-hard [Håstad97] 2 MAX-CUT: If exists cut with (1- ± ) fraction of edges, efficiently cut at least (1-c  ± ) fraction of edges.

1 2  (n) n O(1) exponential polynomial running- time Exponential hardness: time 2 n 1-o(1) Sharp threshold: Drop at ½+o(1) Assuming it takes 2  (n) time to solve 3SAT exactly on size n inputs. 1/2 The Complexity of Approximating 3XOR [=3LIN(GF(2))] [ AroraSafra92, AroraLundMotwaniSudanSzegedy92, Raz94, BellareGoldreichSudan95, Håstad97, MRaz08 ] Approx. factor 3

23 OR [FG95] 7/8 XOR [GW92] 1/2 AND [FG95] 1/2 Best Approximations Known For 2CSP vs. 3CSP any better approximation: NP-hrd [Håstad97] Any better approximation: NP-hard assuming the Unique Games Conjecture [K02,KKMO04,R08] 4 Hard time showing easy problems are hard!

Proving Hardness of Constraint Satisfaction Problems 5 The Bellare-Goldreich-Sudan-Håstad paradigm 2CSP(GF(q)) desired problem composition with long- code/dictator code [Khot02]: Start with 2LIN(GF(q)). The Unique Games Conjecture: 2LIN(GF(q)) is extremely hard to approximate.

The Unique Games Conjecture (UGC) - The Hardness of 2LIN Over Large Alphabets 6 Unique Games Conjecture [Khot02, formulation by KKMO04] : For all ,δ>0, for sufficiently large q, given a set of linear equations of the form x i – x j = c (mod q) where  1-δ fraction of the equations can be satisfied, it is NP-hard to find an assignment that satisfies   fraction of the equations. [Raghavendra08]: Assuming the Unique Games Conjecture, the best approximation for all CSPs is obtained by rounding a natural SDP.

Alphabet/Hardness Rules of Thumb: As alphabet gets larger, problem gets harder. Important Exception: problem can (but not necessarily) become easy if alphabet is R : linear/semidefinite programming can be used. Typically, Easy, if allow fractional solutions Hard, if encode large discrete alphabets (e.g., exact 3LIN( R ) [GR07]) 7

A natural optimization problem: find a balanced assignment for real homogeneous linear equations: SDP-based algorithm: Best approximation known obtained by rounding natural SDP (when 1-  equations exactly satisfiable, get margin O(  ) for  (1) of equations), – Unlike GF(q): Same for 2 or 3 variables per equation! We show NP-hardness for 3 vars per equation! Approach to proving the Unique Games Conj. 8 x 15 - x x 37 = 0 x 1 - 2x 3 + x 89 = 0 ...  Work with Subhash Khot, 2010: Margin of equation ax+by+cz = 0 is |ax+by+cz|

Main Theorem Theorem [KM10]: For any ± >0, it is NP-hard, given a system of linear equations over reals where (1- ± ) fraction can be exactly satisfied by a balanced assignment, to find an assignment where 0.99 fraction of the equations have margin at most ¢  ±. Tight: Efficient algorithm gives margin O(  ) for  (1) of equations. Blow up: Reduction from SAT has blow-up n  n poly(1/  ), matching the recent result of Arora, Barak, Steurer. 9

Approach to Proving The Unique Games Conjecture 1.Show that approximate 3LIN over the reals is NP-hard. 2.Show that approximate 2LIN over the reals is NP-hard, possibly by reduction from 3LIN. 3.Deduce the Unique Games Conjecture using parallel repetition. 10

Approximation Algorithm Min E[| ax + by + cz | 2 ] s.t. assignment is balanced Analysis & Rounding. Assume ( 1-  ) equations exactly satisfiable. Then the SDP minimum is at most O(  ). Solve the SDP to get vector assignment. Pick random Gaussian ³, get real assignment x i = with similar value. The margin is  (  ) for constant fraction of equations. SDP 11

Hardness of Approximately Solving Real Linear Equations with Three Variables Per Equation 12

Proving Hardness of Approximate Real Equations 1.Dictatorship test over reals. 2.Adapting the paradigm 13 The Bellare-Goldreich-Sudan-Håstad paradigm 2CSP(GF(q)) desired problem composition with long- code/dictator code dictatorship test for desired problem

Real Functions Variables correspond to points in R n. Assignments correspond to functions R n  R. We consider the Gaussian distribution over R n, where each coordinate has mean 0 and variance 1. Fact: If x,y are independent Gaussians, then px+qy is Gaussian where p 2 +q 2 = 1. 14

F F Dictator Testing Tester Function F: R n  R. Two possibilities for F. A tester queries three positions x,y,z  R n, tests aF(x)+bF(y)+cF(z) = 0 ? If dictator, equation satisfied exactly with probability  1- . If not approximated by linear junta, there is a margin    with probability 

Noise Sensitivity Approach to Dictator Testing Pick Gaussian x 2 R n. Perturb x by re-sampling each coordinate with probability ±. Obtain x’. Check F(x)=F(x’). 16

Problem with Noise Sensitivity Approach For F half-space, – With probability 1-c  ±, equality holds. – With probability c  ±, constant margin. We want: with probability ¸ 0.01, margin ¸  ±. 17

Linear Non-Juntas Are Noise-Sensitive Observation: If F =  i 2I a i x i for | I | À 1/ ± 2, then (F(x) – F(x’)) is Gaussian with variance ¸  ± . Hence |F(x)-F(x’)| is ¸  ± with prob ¸ Basic Idea: test linearity and noise sensitivity. 18

Hermite Analysis – Fourier Anlysis for Real Functions Fact: For any F: R n  R with bounded |f| 2, can write F =  c i 1,…, i n ¢ H i 1 (x 1 ) ¢ … ¢ H i n (x n ), where H d is a polynomial (“Hermite polynomial”) of degree d. Notation: F · 1 is the linear part of F, and F >1 (x) is its non-linear part. 19

Linearity Testing Pick Gaussian x,y 2 R n, Pick p 2 [0,1]; set q=  (1-p 2 ). Check F(px+qy)=pF(x)+qF(y). Lemma: |F-F · 1 | 2 2 · E[Margin 2 ]. Proof: Via Hermite analysis. 20 Equation depends on three variables!

Decompose into linear and non-linear part: E x |F(x)-F(x’)| 2 = E x |F · 1 (x)-F · 1 (x’)| 2 + E x |F >1 (x)-F >1 (x’)| 2 Non junta ) with prob 0.1 margin ¸ 0.1  Linearity test ) | F >1 (x )| 2 ·  Cancellation problem: On average | F >1 (x )- F >1 (x ’)| · , but it can be much larger when |F · 1 (x)-F · 1 (x’)| is large and cancel |F · 1 (x)-F · 1 (x’)| ! 21

Cancelations Don’t Arise! coordinate-wise perturbation x » c x’: x’ is obtained from x by re-sampling with prob ±. random perturbation x » r x’: x’ is obtained from x as px+qy for p=(1- ± ), p 2 +q 2 =1 Cancelation: P x » c x’ [|F >1 (x)-F >1 (x’)| 2 ¸ 0.1  ] ¸ 0.1 By linearity testing: P x » r x’ [|F >1 (x)-F >1 (x’)| 2 · 0.01  ] ¸ 0.99 We’ll show that they cannot co-exist! 22

Cancelations Don’t Arise! coordinate-wise perturbation x » c x’: x’ is obtained from x by re-sampling with prob ±. random perturbation x » r x’: x’ is obtained from x as px+qy for p=(1- ± ), p 2 +q 2 =1 Cancelation: P x » c x’ [|F >1 (x)-F >1 (x’)| 2 ¸ 0.1  ] ¸ 0.1 By linearity testing: P x » r x’ [|F >1 (x)-F >1 (x’)| 2 · 0.01  ] ¸ 0.99 We’ll show that they cannot co-exist! 23 Claim: For any G: R n  R, E x » r x’ |G(x)-G(x’)| 2 ¸ E x » c x’ |G(x)-G(x’)| 2.

Main Lemma: From Small Difference to Constant Difference Main Lemma: If for H: R n  R, P x » c x’ [|H(x)-H(x’)| 2 ¸ A ¢ T] ¸ 10 ®, P x » r x’ [|H(x)-H(x’)| 2 · T] ¸ 1- ®, Then, there is Boolean B: R n  {0,1} where P x » c x’ [|B(x)-B(x’)| 2 = 1] ¸ 0.01, P x » r x’ [|B(x)-B(x’)| 2 = 0] ¸ 1-1/A, 24 Proof: Combinatorial, by considering “perturbation graphs”, and constructing an appropriate cut.

Thank you! 25

Proof of Main Lemma Perturbation graphs: Our task: Find a cut in C µ R n that cuts: E c edges of weight ¸ E r edges of weight ·

Cutting The Gaussian Space Claim: There is a distribution D over cuts s.t.: Every edge e 2 E c is cut with prob ¸ 0.1 . Every edge e 2 E r is cut with prob · 0.01  -1/21/2 H(x)H(x’) 27

From Small Cuts To Large Cuts In previous lemma, cut only ¼  weight. Want to cut constant weight. 28

Larger Cut Let m=11/ . Sample from D cuts C 1,…,C m. Pick random Iµ [m]. Let C =  i 2I C i. Claim: Edge e 2 E c is cut with prob ¸ 0.5 ¢ (1-1/e 2 ). Edge e 2 E r is cut with prob  0.01  ¢ 10/  0.1  Corr: A cut with 0.2 ¢ 0.1 weight of E c and 0.8 ¢ 0.99 weight of E r. 29