S.Safra some slides borrowed from Dana Moshkovits

Slides:



Advertisements
Similar presentations
Inapproximability of Hypergraph Vertex-Cover. A k-uniform hypergraph H= : V – a set of vertices E - a collection of k-element subsets of V Example: k=3.
Advertisements

Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
MaxClique Inapproximability Seminar on HARDNESS OF APPROXIMATION PROBLEMS by Dr. Irit Dinur Presented by Rica Gonen.
1/17 Optimal Long Test with One Free Bit Nikhil Bansal (IBM) Subhash Khot (NYU)
Complexity ©D.Moshkovits 1 Hardness of Approximation.
Complexity ©D.Moshkovitz 1 Introduction. Complexity ©D.Moshkovitz 2 Introduction Objectives: –To introduce some of the basic concepts of complexity theory.
1 Tight Hardness Results for Some Approximation Problems [Raz,Håstad,...] Adi Akavia Dana Moshkovitz & Ricky Rosen S. Safra.
Fourier Analysis, Projections, Influences, Juntas, Etc…
NP-Complete Problems Reading Material: Chapter 10 Sections 1, 2, 3, and 4 only.
1 Tight Hardness Results for Some Approximation Problems [mostly Håstad] Adi Akavia Dana Moshkovitz S. Safra.
1 INTRODUCTION NP, NP-hardness Approximation PCP.
CSE 421 Algorithms Richard Anderson Lecture 27 NP Completeness.
The Importance of Being Biased Irit Dinur S. Safra (some slides borrowed from Dana Moshkovitz) Irit Dinur S. Safra (some slides borrowed from Dana Moshkovitz)
Fourier Analysis of Boolean Functions Juntas, Projections, Influences Etc.
Complexity 1 Hardness of Approximation. Complexity 2 Introduction Objectives: –To show several approximation problems are NP-hard Overview: –Reminder:
Complexity ©D.Moshkovits 1 Hardness of Approximation.
1 Slides by Asaf Shapira & Michael Lewin & Boaz Klartag & Oded Schwartz. Adapted from things beyond us.
1 Joint work with Shmuel Safra. 2 Motivation 3 Motivation.
Dana Moshkovitz, MIT Joint work with Subhash Khot, NYU.
Correlation testing for affine invariant properties on Shachar Lovett Institute for Advanced Study Joint with Hamed Hatami (McGill)
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
Nattee Niparnan. Easy & Hard Problem What is “difficulty” of problem? Difficult for computer scientist to derive algorithm for the problem? Difficult.
Week 10Complexity of Algorithms1 Hard Computational Problems Some computational problems are hard Despite a numerous attempts we do not know any efficient.
1/19 Minimizing weighted completion time with precedence constraints Nikhil Bansal (IBM) Subhash Khot (NYU)
Non-Approximability Results. Summary -Gap technique -Examples: MINIMUM GRAPH COLORING, MINIMUM TSP, MINIMUM BIN PACKING -The PCP theorem -Application:
NP-Complete problems.
CS 3343: Analysis of Algorithms Lecture 25: P and NP Some slides courtesy of Carola Wenk.
CSE 589 Part V One of the symptoms of an approaching nervous breakdown is the belief that one’s work is terribly important. Bertrand Russell.
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
CS6045: Advanced Algorithms NP Completeness. NP-Completeness Some problems are intractable: as they grow large, we are unable to solve them in reasonable.
1 2 Introduction In this lecture we’ll cover: Definition of PCP Prove some classical hardness of approximation results Review some recent ones.
NP Completeness Piyush Kumar. Today Reductions Proving Lower Bounds revisited Decision and Optimization Problems SAT and 3-SAT P Vs NP Dealing with NP-Complete.
Lecture. Today Problem set 9 out (due next Thursday) Topics: –Complexity Theory –Optimization versus Decision Problems –P and NP –Efficient Verification.
CSE 421 Algorithms Richard Anderson Lecture 27 NP-Completeness Proofs.
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
COSC 3101A - Design and Analysis of Algorithms 14 NP-Completeness.
The NP class. NP-completeness Lecture2. The NP-class The NP class is a class that contains all the problems that can be decided by a Non-Deterministic.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
ICS 353: Design and Analysis of Algorithms NP-Complete Problems King Fahd University of Petroleum & Minerals Information & Computer Science Department.
Tight Hardness Results for Some Approximation Problems [Håstad]
The Theory of NP-Completeness
The NP class. NP-completeness
More NP-Complete and NP-hard Problems
P & NP.
Chapter 10 NP-Complete Problems.
Richard Anderson Lectures NP-Completeness
Richard Anderson Lecture 26 NP-Completeness
Richard Anderson Lecture 26 NP-Completeness
Complexity Introduction Complexity ©D.Moshkovitz.
NP-Completeness Yin Tat Lee
Where Can We Draw The Line?
ICS 353: Design and Analysis of Algorithms
Parameterised Complexity
Richard Anderson Lecture 25 NP-Completeness
Noise-Insensitive Boolean-Functions are Juntas
Richard Anderson Lecture 28 NP-Completeness
Introduction to PCP and Hardness of Approximation
Chapter 34: NP-Completeness
Hardness of Approximation
CS 3343: Analysis of Algorithms
NP-Completeness Yin Tat Lee
Interactive Proofs Adapted from Oded Goldreich’s course lecture notes.
The Theory of NP-Completeness
Complexity Theory in Practice
Instructor: Aaron Roth
Instructor: Aaron Roth
The Importance of Being Biased
CHAPTER 7 Time complexity
Lecture 23 NP-Hard Problems
Presentation transcript:

S.Safra some slides borrowed from Dana Moshkovits Probabilistically Checkable Proofs and Hardness of Approximation S.Safra some slides borrowed from Dana Moshkovits

The Crazy Tea Party Problem To seat all guests at a round table, so people who sit an adjacent seats like each other. John Mary Bob Jane Alice 

Solution for the Example Problem To seat all guests at a round table, so people who sit an adjacent seats like each other. John Jane Mary Alice Bob

Naive Algorithm For each ordering of the guests around the table Verify each guest likes the guest sitting in the next seat.

How Much Time Should This Take? (worse case) guests steps n (n-1)! 5 24 say our computer is capable of 1010 instructions per second, this will still take  3·10138 years! 15 87178291200 100 9·10155

Problem Plan a trip that visits every site exactly once. Tours Problem Plan a trip that visits every site exactly once.

Solution for the Example Problem Plan a trip that visits every site exactly once.

Is a Problem Tractable? YES! And here’s an efficient algorithm for it NO! and I can prove it and what if neither is the case?

Growth Rate: Sketch n! =2O(n lg n) time 2n n2 10n input length

The World According to Complexity reasonable unreasonable polynomial  nO(1) exponential  2nO(1)

Could one be Fundamentally Harder than the Other? Seating Tour

Relations Between Problems Assuming an efficient procedure for problem A, there is an efficient procedure for problem B B cannot be radically harder than A

Reductions B A p B cannot be radically harder than A In other words: A is at least as hard as B

Which One is Harder? ? Seating Tour

“directly reachable from…” Reduce Tour to Seating First Observation: The problems aren’t so different site guest “directly reachable from…” “liked by…”

Reduce Tour to Seating Second Observation: Completing the circle Let’s invite to our party a very popular guest, i.e one who can sit next to everybody else.

Reduce Tour to Seating If there is a tour, there is also a way to seat all the imagined guests around the table. popular guest . . . . . .

Reduce Tour to Seating If there is a seating, we can easily find a tour path (no tour, no seating). popular guest . . . . . .

Bottom Line The seating problem is at least as hard as the tour problem

What have we shown? Although we couldn’t come up with an efficient algorithm for the problems Nor to prove they don’t have one, We managed to show a very powerful claim regarding the relation between their hardness

Furthermore Interestingly, we can also reduce the seating problem to the tour problem. Moreover, there is a whole class of problems, which can be pair-wise efficiently reduced to each other.

NPC NPC  ? exponential algorithms efficient algorithms Contains thousands of distinct problem NPC each reducible to all others  exponential algorithms efficient algorithms ?

How can Studying P vs NP Make You a Millionaire? This is the most fundamental open question of computer science. Resolving it would grant the solver a great honor … as well as substantial fortune… www.claymath.org/prizeproblems/pvsnp.htm Huge philosophical implications: No need for human ingenuity! No need for mathematicians!!!

Constraints Satisfaction Def Constraints Satisfaction Problem (CSP): Instance: Constraints: A set of constraints  = { 1, …, l } over two sets of variables, X of range RX and Y of range RY Determinate: each constraint determines the value of a variable yY according to the value of some xX xy : RX  RY , satisfied if xy(x)=y Uniform: each xX appears in dX of , and each yY appears in dY of , for some global dX and dy Optimize: Define () = maximum, over all assignments to X and Y A: X RX; Y RY of the fraction of  satisfied

Cook’s Characterization of NP Thm: It is NP-hard to distinguish between () = 1 () < 1 For any language L in NP CSP testing membership in L can be reduced to...

Showing hardness From now on, to show a problem NP-hard, we merely need to reduce CSP to it. any NP problem can be reduced to... CSP can be reduced to... Cook’s Thm new, hard problem will imply the new problem is NP-hard

Max Independent-Set Instance: A graph G=(V,E) and a threshold k. Problem: To decide if there is a set of vertices I={v1,...,vk}V, s.t. for any u,vI: (u,v)E.

Max I.S. is NP-hard Proof: We’ll show CSPp Max I.S. ≤p k

The reduction: Co-Partite Graph G comprise k=|X| cliques of size |RX| - a vertex for each plausible assignment to x: k An edge: two assignments that determine a different value to same y E  {(<i,j1>, <i,j2>) | iM, j1≠j2 RX}

Proof of Correctness An I.S. of size k must contain exactly one vertex in every clique. k A satisfying assignment implies an I.S. of size k An I.S. of size k corresponds to a consistent, satisfying assignment

Generalized Tour Problem Add prices to the roads of the tour problem Ask for the least costly tour $3 $17 $13 $8 $19 $10 $13 $12

Approximation How about approximating the optimal tour? I.e – finding a tour which costs, say, no more than twice as much as the least costly. $3 $17 $13 $8 $19 $10 $13 $12

Hardness of Approximation

Promise Problems Sometimes you can promise something about the input It doesn’t matter what you say for unfeasible inputs I know my graph has clique of size n/4! Does it have a clique of size n/2?

Promise Problems & Approximation We’ll see promise problems of a certain type, called gap problems, can be utilized to prove hardness of approximation.

Gap Problems (Max Version) Instance: … Problem: to distinguish between the following two cases: The maximal solution  B The maximal solution ≤ A YES NO

Idea We’ve shown “standard” problems are NP-hard by reductions from CSP. We want to prove gap-problems are NP-hard Why won’t we prove some canonical gap-problem is NP-hard and reduce from it? If a reduction reduces one gap-problem to another we refer to it as gap-preserving

Gap-CSP[] YES NO Instance: Same as CSP Problem: to distinguish between the following two cases: There exists an assignment that satisfies all constraints. No assignment can satisfy more than  of the constraints. YES NO

Gap-CSP[] is NP-hard, as long as |RX|,|RY| ≥ -O(1) PCP (Without Proof) Theorem [FGLSS, AS, ALMSS]: For any >0, Gap-CSP[] is NP-hard, as long as |RX|,|RY| ≥ -O(1)

Why Is It Called PCP? (Probabilistically Checkable Proofs) CSP has a polynomial membership proof checkable in polynomial time. Prove it! This assignment satisfies it! My formula is satisfiable! x1 x2 x3 x4 x5 x6 x7 x8 yn-3 yn-2 yn-1 yn . . .

Why Is It Called PCP? (Probabilistically Checkable Proofs) …Now our verifier has to check the assignment satisfies all constraints…

Why Is It Called PCP? (Probabilistically Checkable Proofs) In a NO instance of gap-CSP, 1- of the constraints are not satisfied! While for gap-CSP the verifier would be right with high probability, even by: pick at random a constant number of constraints and check only those

Why Is It Called PCP? (Probabilistically Checkable Proofs) Since gap-CSP is NP-hard, All NP problems have probabilistically checkable proofs.

Hardness of Approximation Do the reductions we’ve seen also work for the gap versions (i.e approximation preserving)? We’ll revisit the Max I.S. example.

The same Max I.S. Reduction An I.S. of size k must contain exactly one vertex in every part. k A satisfying assignment implies an I.S. of size k An I.S. of size k corresponds to a consistent assignment satisfying  of 

Corollary Theorem: for any >0, Independent-set is hard to approximate to within any constant factor

Chromatic Number Instance: a graph G=(V,E). Problem: To minimize k, so that there exists a function f:V{1,…,k}, for which (u,v)E  f(u)f(v) skip

Observation: Each color class is an independent set Chromatic Number Observation: Each color class is an independent set

Clique Cover Number (CCN) Instance: a graph G=(V,E). Problem: To minimize k, so that there exists a function f:V{1,…,k}, for which (u,v)E  f(u)=f(v)

Clique Cover Number (CCN)

Observation Claim: The CCN problem on graph G is the CHROMATIC-NUMBER problem on the complement graph Gc.

Reduction Idea . . CLIQUE CCN G G’ m same under cyclic shift clique preserving q

Correctness Given such transformation: MAX-CLIQUE(G) = m  CCN(G’) = q

T is unique for triplets Transformation T:V[q] for any v1,v2,v3,v4,v5,v6, T(v1)+T(v2)+T(v3) T(v4)+T(v5)+T(v6) (mod q) {v1,v2,v3}={v4,v5,v6} T is unique for triplets

Observations Such T is unique for pairs and for single vertices as well: If T(x)+T(u)=T(v)+T(w) (mod q), then {x,u}={v,w} If T(x)=T(y) (mod q), then x=y

Using the Transformation vi vj CLIQUE T(vj)=4 T(vi)=1 CCN 0 1 2 3 4 … (q-1)

Completing the CCN Graph Construction T(s) (s,t)ECLIQUE  (T(s),T(t))ECCN T(t)

Completing the CCN Graph Construction Close the set of edges under shift: For every (x,y)E, if x’-y’=x-y (mod q), then (x’,y’)E T(s) T(t)

First Observation: This edge comes only from (s,t) Edge Origin Unique T(s) First Observation: This edge comes only from (s,t) T(t)

Second Observation: A triangle only comes from a triangle Triangle Consistency Second Observation: A triangle only comes from a triangle

Clique Preservation Corollary: {c1,…,ck} is a clique in the CCN graph iff {T(c1),…,T(ck)} is a clique in the CLIQUE graph.

What Remains? It remains to show how to construct the transformation T in polynomial time.

Corollaries Theorem: CCN is NP-hard to approximate within any constant factor. Theorem: CHROMATIC-NUMBER is NP-hard to approximate within any constant factor.

Max-E3-Lin-2 Def: Max-E3-Lin-2 Instance: a system of linear equations L = { E1, …, En } over Z2 each equation of exactly 3 variables (whose sum is required to equal either 0 or 1) Problem: Compute (L)

Main Theorem Thm [Hastad]: gap-Max-E3-Lin-2(1-, ½+) is NP-hard. That is, for every constant >0 it is NP-hard to distinguish between the case 1- of the equations are satisfiable and the case ½+ are. [ It is therefore NP-Hard to approximate Max-E3-Lin-2 to within 2- constant >0]

This bound is Tight! A random assignment satisfies half of the equations. Deciding whether a set of linear equations have a common solution is in P (Gaussian elimination).

Proof Outline The proof proceeds with a reduction from gap-CSP[], known to be NP-hard for any constant >0 Given such an instance , the proof shows a poly-time construction, of an instance L of Max-E3-Lin-2 s.t. () = 1  (L) ≥ 1 - L () <   (L) ≤ ½ + L Main Idea: Replace every x and every y with a set of variables representing a binary-code of their assigned values. Then, test consistency within encoding and any xy using linear equations over 3 bits

Long-Code of R One bit for every subset of R

Long-Code of R One bit for every subset of R to encode an element eR 1 1 1

In fact use a “folded” long-code, s.t. f(F)=1-f([n]\F) The Variables of L Consider an instance  of CSP[], for small constant  (to be fixed later) L has 2 types of variables: a variable z[y,F] for every variable yY and a subset F  P[Ry] a variable z[x,F] for every variable xX and a subset F  P[RX] In fact use a “folded” long-code, s.t. f(F)=1-f([n]\F)

Linearity of a Legal-Encoding An Boolean function f: P[R]  Z2, if legal long-code-word , is a linear-function, that is, for every F, G  P[R]: f(F) + f(G)  f(FG) where FG  P[R] is the symmetric difference of F and G Unfortunately, any linear function (a sum of a subset of variables) will pass this test

The Distribution  Def: denote by  the biased, product distribution over P[RX], which assigns probability to a subset H as follows: Independently, for each aRX, let aH with probability 1- aH with probability  One should think of  as a multiset of subsets in which every subset H appears with the appropriate probability

The Linear Equations L‘s linear-equations are the union, over all , of the following set of equations:  F P[RY], G P[RX] and H  denote F*= xy-1(F) z[y,F] + z[x, G]  z[x, F*  G  H]

Correctness of Reduction Prop: if () = 1 then (L) = 1- Proof: let A be a satisfying assignment to . Assign all L ‘s variables according to the legal-encoding of A’s values. A linear equation of L, corresponding to xy,F,G,H, would be unsatisfied exactly if A(x)H, which occurs with probability  over the choice of H. LLC-Lemma: (L) = ½+/2  () > 42 Note: independent of ! (Later we use that fact to set  small enough for our needs).  = 2(L) -1

Denoting an Assignment to L Given an assignment AL to L’s variables: For any xX, denote by fx : P[RX]  {-1, 1} the function comprising the values AL assigns to z[x,*] (corresponding to the long-code of the value assigned to x) For any yY, denote by fy : P[RY]  {-1, 1} the function comprising the values AL assigns to z[y,*] (corresponding to the long-code of the value assigned to y) Replacing 1 by -1 and 0 by 1

Distributional Assignments Consider a CSP instance  Let (R) be the set of all distributions over R Def: A distributional-assignment to  is A: X  (RX); Y  (RX) Denote by () the maximum over distributional-assignments A of the average probability for    to be satisfied, if variables’ values are chosen according to A Clearly ()  (). Moreover Prop: ()  ()

The Distributional-Assignment A Def: Let A be a distributional-assignment to  according to the following random processes: For any variable xX Choose a subset SRX with probability Uniformly choose a random aS. For any variable yY Choose a subset SRY with probability Uniformly choose a random bS. For such functions, the squares of the coefficients constitute a distribution

odd(xy(S)) = {b| #{aS| xy(a) = b} is odd} What’s to do: Show that AL‘s expected success on xy is > 42 in two steps: First show that AL‘s success probability, for any xy Then show that value to be  42

Claim 1 Claim 1: AL‘s success probability, for any xy Proof: That success probability is Now, taking the sum for only the cases in which Sy=odd(xy(Sx)), results in the claimed inequality.

High Success Probability

Related work Thm (Friedgut): a Boolean function f with small average-sensitivity is an [,j]-junta Thm (Bourgain): a Boolean function f with small high-frequency weight is an [,j]-junta Thm (Kindler&Safra): a Boolean function f with small high-frequency weight in a p-biased measure is an [,j]-junta Corollary: a Boolean function f with small noise-sensitivity is an [,j]-junta [Dinur, S] Showing Vertex-Cover hard to approximate to within 10 5 – 21 Parameters: average-sensitivity [BL,KKL,F]; high-frequency weight [H,B], noise-sensitivity [BKS]

Boolean Functions and Juntas A Boolean function Def: f is a j-Junta if there exists J[n] where |J|≤ j, and s.t. for every x f(x) = f(x  J) f is (, j)-Junta if  j-Junta f’ s.t.

Motivation – Testing Long-code Def (a long-code test): given a code-word w, probe it in a constant number of entries, and accept w.h.p if w is a monotone dictatorship reject w.h.p if w is not close to any monotone dictatorship

Motivation – Testing Long-code Def(a long-code list-test): given a code-word w, probe it in a constant number of entries, and accept w.h.p if w is a monotone dictatorship, reject w.h.p if  a Junta J[n] s.t. f is close to f’ and f’(F)=f’(FJ) for all F Note: a long-code list-test, distinguishes between the case w is a dictatorship, to the case w is far from a junta.

Motivation – Testing Long-code The long-code test, and the long-code list-test are essential tools in proving hardness results. Examples … Hence finding simple sufficient-conditions for a function to be a junta is important.

Noise-Sensitivity Idea: check how the value of f changes when the input is changed not on one, but on several coordinates. [n] I z x

Noise-Sensitivity Def(,p,x[n] ): Let 0<<1, and xP([n]). Then y~,p,x, if y = (x\I) z where I~[n] is a noise subset, and z~ pI is a replacement. Def(-noise-sensitivity): let 0<<1, then Note: deletes a coordinate in x w.p. (1-p), adds a coordinate to x w.p. p. Hence, when p=1/2: equivalent to flipping each coordinate in x w.p. /2. [n] x I z

Noise-Sensitivity – Cont. Advantage: very efficiently testable (using only two queries) by a perturbation-test. Def (perturbation-test): choose x~p, and y~,p,x, check whether f(x)=f(y). The success is proportional to the noise-sensitivity of f. Prop: the -noise-sensitivity is given by

Related Work [Dinur, S] Showing Vertex-Cover hard to approximate to within 10 5 – 21 [Bourgain] Showing a Boolean function with weight <1/k on characters of size larger than k, is close to a junta of size exponential in k ([Kindler, S] similar for biased, product distribution)