CS151 Complexity Theory Lecture 7 April 20, 2004.

Slides:



Advertisements
Similar presentations
Complexity Theory Lecture 6
Advertisements

CS151 Complexity Theory Lecture 3 April 6, CS151 Lecture 32 Introduction A motivating question: Can computers replace mathematicians? L = { (x,
Having Proofs for Incorrectness
Probabilistic algorithms Section 10.2 Giorgi Japaridze Theory of Computability.
Randomized Algorithms Kyomin Jung KAIST Applied Algorithm Lab Jan 12, WSAC
Complexity 26-1 Complexity Andrei Bulatov Interactive Proofs.
1 L is in NP means: There is a language L’ in P and a polynomial p so that L 1 · L 2 means: For some polynomial time computable map r : 8 x: x 2 L 1 iff.
Complexity 18-1 Complexity Andrei Bulatov Probabilistic Algorithms.
CS151 Complexity Theory Lecture 3 April 6, Nondeterminism: introduction A motivating question: Can computers replace mathematicians? L = { (x,
CS151 Complexity Theory Lecture 6 April 15, 2015.
CS151 Complexity Theory Lecture 5 April 13, 2015.
CS21 Decidability and Tractability
CS151 Complexity Theory Lecture 5 April 13, 2004.
Randomized Computation Roni Parshani Orly Margalit Eran Mantzur Avi Mintz
Arithmetic Hardness vs. Randomness Valentine Kabanets SFU.
CS151 Complexity Theory Lecture 7 April 20, 2015.
Submitted by : Estrella Eisenberg Yair Kaufman Ohad Lipsky Riva Gonen Shalom.
Randomized Computation
Complexity 19-1 Complexity Andrei Bulatov More Probabilistic Algorithms.
–Def: A language L is in BPP c,s ( 0  s(n)  c(n)  1,  n  N) if there exists a probabilistic poly-time TM M s.t. : 1.  w  L, Pr[M accepts w]  c(|w|),
CS151 Complexity Theory Lecture 11 May 4, CS151 Lecture 112 Outline Extractors Trevisan’s extractor RL and undirected STCONN.
Derandomizing LOGSPACE Based on a paper by Russell Impagliazo, Noam Nissan and Avi Wigderson Presented by Amir Rosenfeld.
CS151 Complexity Theory Lecture 13 May 11, CS151 Lecture 132 Outline Natural complete problems for PH and PSPACE proof systems interactive proofs.
Chapter 11: Limitations of Algorithmic Power
CS151 Complexity Theory Lecture 15 May 18, CS151 Lecture 152 Outline IP = PSPACE Arthur-Merlin games –classes MA, AM Optimization, Approximation,
CS151 Complexity Theory Lecture 6 April 15, 2004.
Complexity ©D. Moshkovitz 1 And Randomized Computations The Polynomial Hierarchy.
The Polynomial Hierarchy By Moti Meir And Yitzhak Sapir Based on notes from lectures by Oded Goldreich taken by Ronen Mizrahi, and lectures by Ely Porat.
CS151 Complexity Theory Lecture 9 April 27, 2004.
Foundations of Cryptography Lecture 2 Lecturer: Moni Naor.
Ragesh Jaiswal Indian Institute of Technology Delhi Threshold Direct Product Theorems: a survey.
The Complexity of Primality Testing. What is Primality Testing? Testing whether an integer is prime or not. – An integer p is prime if the only integers.
Approximation Algorithms Pages ADVANCED TOPICS IN COMPLEXITY THEORY.
February 18, 2015CS21 Lecture 181 CS21 Decidability and Tractability Lecture 18 February 18, 2015.
Computational Complexity Theory Lecture 2: Reductions, NP-completeness, Cook-Levin theorem Indian Institute of Science.
Theory of Computing Lecture 17 MAS 714 Hartmut Klauck.
CS151 Complexity Theory Lecture 13 May 11, Outline proof systems interactive proofs and their power Arthur-Merlin games.
1 Interactive Proofs proof systems interactive proofs and their power Arthur-Merlin games.
CSCI 2670 Introduction to Theory of Computing November 29, 2005.
1 2 Probabilistic Computations  Extend the notion of “efficient computation” beyond polynomial-time- Turing machines.  We will still consider only.
Computation Model and Complexity Class. 2 An algorithmic process that uses the result of a random draw to make an approximated decision has the ability.
Week 10Complexity of Algorithms1 Hard Computational Problems Some computational problems are hard Despite a numerous attempts we do not know any efficient.
1 Lower Bounds Lower bound: an estimate on a minimum amount of work needed to solve a given problem Examples: b number of comparisons needed to find the.
. CLASSES RP AND ZPP By: SARIKA PAMMI. CONTENTS:  INTRODUCTION  RP  FACTS ABOUT RP  MONTE CARLO ALGORITHM  CO-RP  ZPP  FACTS ABOUT ZPP  RELATION.
1 The Theory of NP-Completeness 2 Cook ’ s Theorem (1971) Prof. Cook Toronto U. Receiving Turing Award (1982) Discussing difficult problems: worst case.
CS151 Complexity Theory Lecture 12 May 6, QSAT is PSPACE-complete Theorem: QSAT is PSPACE-complete. Proof: 8 x 1 9 x 2 8 x 3 … Qx n φ(x 1, x 2,
Umans Complexity Theory Lectures Lecture 1a: Problems and Languages.
Fall 2013 CMU CS Computational Complexity Lecture 7 Alternating Quantifiers, PH, PSPACE. These slides are mostly a resequencing of Chris Umans’
Fall 2013 CMU CS Computational Complexity Lectures 8-9 Randomness, communication, complexity of unique solutions These slides are mostly a resequencing.
Randomization Carmella Kroitoru Seminar on Communication Complexity.
Umans Complexity Theory Lectures Lecture 7b: Randomization in Communication Complexity.
Computability & Complexity II Chris Umans Caltech.
CS151 Complexity Theory Lecture 16 May 20, The outer verifier Theorem: NP  PCP[log n, polylog n] Proof (first steps): –define: Polynomial Constraint.
Pseudo-random generators Talk for Amnon ’ s seminar.
Complexity 24-1 Complexity Andrei Bulatov Interactive Proofs.
CSCI 2670 Introduction to Theory of Computing December 2, 2004.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lectures
The NP class. NP-completeness Lecture2. The NP-class The NP class is a class that contains all the problems that can be decided by a Non-Deterministic.
Pseudo-randomness. Randomized complexity classes model: probabilistic Turing Machine –deterministic TM with additional read-only tape containing “coin.
PROBABILITY AND COMPUTING RANDOMIZED ALGORITHMS AND PROBABILISTIC ANALYSIS CHAPTER 1 IWAMA and ITO Lab. M1 Sakaidani Hikaru 1.
Umans Complexity Theory Lectures
Probabilistic Algorithms
Umans Complexity Theory Lectures
CS21 Decidability and Tractability
Umans Complexity Theory Lectures
CS21 Decidability and Tractability
The Polynomial Hierarchy
Umans Complexity Theory Lectures
CS151 Complexity Theory Lecture 8 April 26, 2019.
CS151 Complexity Theory Lecture 7 April 23, 2019.
Presentation transcript:

CS151 Complexity Theory Lecture 7 April 20, 2004

CS151 Lecture 72 Outline 3 examples of the power of randomness –communication complexity –polynomial identity testing –complexity of finding unique solutions randomized complexity classes Adelman’s Theorem

April 20, 2004CS151 Lecture 73 Communication complexity Goal: compute f(x, y) while communicating as few bits as possible between Alice and Bob count number of bits exchanged (computation free) at each step: one party sends bits that are a function of held input and received bits so far two parties: Alice and Bob function f:{0,1} n x {0,1} n  {0,1} Alice holds x  {0,1} n ; Bob holds y  {0,1} n

April 20, 2004CS151 Lecture 74 Communication complexity simple function (equality): EQ(x, y) = 1 iff x = y simple protocol: –Alice sends x to Bob (n bits) –Bob sends EQ(x, y) to Alice (1 bit) –total: n + 1 bits –(works for any predicate f)

April 20, 2004CS151 Lecture 75 Communication complexity Can we do better? –deterministic protocol? –probabilistic protocol? at each step: one party sends bits that are a function of held input and received bits so far and the result of some coin tosses required to output f(x, y) with high probability over all coin tosses

April 20, 2004CS151 Lecture 76 Communication complexity Theorem: no deterministic protocol can compute EQ(x, y) while exchanging fewer than n+1 bits. Proof: –“input matrix”: X = {0,1} n Y = {0,1} n f(x,y)

April 20, 2004CS151 Lecture 77 Communication complexity –assume without loss of generality 1 bit sent at a time –A sends 1 bit depending only on x: X = {0,1} n Y = {0,1} n inputs x causing A to send 1 inputs x causing A to send 0

April 20, 2004CS151 Lecture 78 Communication complexity –B sends 1 bit depending only on y and received bit: X = {0,1} n Y = {0,1} n inputs y causing B to send 1 inputs y causing B to send 0

April 20, 2004CS151 Lecture 79 Communication complexity –at end of protocol involving k bits of communication, matrix is partitioned into at most 2 k combinatorial rectangles –bits sent in protocol are the same for every input (x, y) in given rectangle –conclude: f(x,y) must be constant on each rectangle

April 20, 2004CS151 Lecture 710 Communication complexity –any partition into combinatorial rectangles with constant f(x,y) must have 2 n + 1 rectangles –protocol that exchanges ≤ n bits can only create 2 n rectangles, so must exchange at least n+1 bits. X = {0,1} n Y = {0,1} n Matrix for EQ:

April 20, 2004CS151 Lecture 711 Communication complexity protocol for EQ employing randomness? –Alice picks random prime p in {1...4n 2 }, sends: p (x mod p) –Bob sends: (y mod p) –players output 1 if and only if: (x mod p) = (y mod p)

April 20, 2004CS151 Lecture 712 Communication complexity –O(log n) bits exchanged –if x = y, always correct –if x ≠ y, incorrect if and only if: p divides |x – y| –# primes in range is ≥ 2n –# primes dividing |x – y| is ≤ n –probability incorrect ≤ 1/2 Randomness gives an exponential advantage!!

April 20, 2004CS151 Lecture 713 Polynomial identity testing Given: polynomial p(x 1, x 2, …, x n ) over field F Is p identically zero? –i.e., is p(x) = 0 for all x  F n –(assume |F| larger than degree…) “polynomial identity testing” because given two polynomials p, q, we can check the identity p  q by checking if (p – q)  0

April 20, 2004CS151 Lecture 714 Polynomial identity testing polynomial p(x 1, x 2, …, x n ) given as arithmetic circuit: + * x1x1 x2x2 * +- x3x3 …xnxn * multiplication (fan-in 2) addition (fan-in 2) negation (fan-in 1)

April 20, 2004CS151 Lecture 715 Polynomial identity testing try all |F| n inputs? –may be exponentially many multiply out symbolically, check that all coefficients are zero? –may be exponentially many coefficients can randomness help? –i.e., flip coins, allow small probability of wrong answer

April 20, 2004CS151 Lecture 716 Polynomial identity testing Lemma (Schwartz-Zippel): Let p(x 1, x 2, …, x n ) be a total degree d polynomial over a field F and let S be any subset of F. Then if p is not identically 0, Pr r 1, r 2,…, r n  S [ p(r 1, r 2, …, r n ) = 0] ≤ d/|S|.

April 20, 2004CS151 Lecture 717 Polynomial identity testing Proof: –induction on number of variables n –base case: n = 1, p is univariate polynomial of degree at most d –at most d roots, so Pr [ p(r 1 ) = 0] ≤ d/|S|

April 20, 2004CS151 Lecture 718 Polynomial identity testing –write p(x 1, x 2, …, x n ) as p(x 1, x 2, …, x n ) = Σ i (x 1 ) i p i (x 2, …, x n ) –k = max. i for which p i (x 2, …, x n ) not id. zero –by induction hypothesis: Pr [ p k (r 2, …, r n ) = 0] ≤ (d-k)/|S| –whenever p k (r 2, …, r n ) ≠ 0, p(x 1, r 2, …, r n ) is a univariate polynomial of degree k Pr[p(r 1,r 2,…,r n )=0 | p k (r 2,…,r n ) ≠ 0] ≤ k/|S|

April 20, 2004CS151 Lecture 719 Polynomial identity testing Pr [ p k (r 2, …, r n ) = 0] ≤ (d-k)/|S| Pr[p(r 1,r 2,…,r n )=0 | p k (r 2,…,r n ) ≠ 0] ≤ k/|S| –conclude: Pr[ p(r 1, …, r n ) = 0] ≤ (d-k)/|S| + k/|S| = d/|S| –Note: can add these probabilities because Pr[E 1 ] = Pr[E 1 |E 2 ]Pr[E 2 ] + Pr[E 1 |  E 2 ]Pr[  E 2 ] ≤ Pr[E 2 ] + Pr[E 1 |  E 2 ]

April 20, 2004CS151 Lecture 720 Polynomial identity testing Given: polynomial p(x 1, x 2, …, x n ) over field F Is p identically zero? Note: degree d is at most the size of input + * x1x1 x2x2 * +- x3x3 …xnxn *

April 20, 2004CS151 Lecture 721 Polynomial identity testing randomized algorithm: pick a subset S  F of size 2d –pick r 1, r 2, …, r n from S uniformly at random –if p(r 1, r 2, …, r n ) = 0, answer “yes” –if p(r 1, r 2, …, r n ) ≠ 0, answer “no” if p identically zero, never wrong if not, Schwartz-Zippel ensures probability of error at most ½

April 20, 2004CS151 Lecture 722 Unique solutions a positive instance of SAT may have many satisfying assignments maybe the difficulty comes from not knowing which to “work on” if we knew # satisfying assignments was 1 or 0, could we zero in on the 1 efficiently?

April 20, 2004CS151 Lecture 723 Unique solutions Question: given polynomial-time algorithm that works on SAT instances with at most 1 satisfying assignment, can we solve general SAT instances efficiently? Answer: yes –but only if “efficiently” allows randomness

April 20, 2004CS151 Lecture 724 Unique solutions Theorem (Valiant-Vazirani): there is a randomized poly-time procedure that given a 3-CNF formula φ(x 1, x 2, …, x n ) outputs a 3-CNF formula φ’ such that –if φ is not satisfiable then φ’ is not satisfiable –if φ is satisfiable then with probability at least 1/(8n) φ’ has exactly one satisfying assignment

April 20, 2004CS151 Lecture 725 Unique solutions Proof: –given subset S  {1, 2, …, n}, there exists a 3-CNF formula θ S on x 1, x 2, …, x n and additional variables such that: |θ S | = O(n) θ S is satisfiable iff an even number of variables in {x i } i  S are true for each such setting of the x i variables, this satisfying assignment is unique not difficult; details omitted

April 20, 2004CS151 Lecture 726 Unique solutions –set φ 0 = φ –for i = 1, 2, …, n pick random subset S i set φ i = φ i-1  θ S i –output random φ i –T = set of satisfying assignments for φ –Claim: if |T| > 0, then Pr k  {0,1,2,…,n-1} [2 k ≤ |T| ≤ 2 k+1 ] ≥ 1/n

April 20, 2004CS151 Lecture 727 Unique solutions Claim: if 2 k ≤ |T| ≤ 2 k+1, then the probability φ k+2 has exactly one satisfying assignment is ≥ 1/8 –fix t  T –Pr[t “agrees with” t’ on S i ] = ½ –Pr[t agrees with t’ on S 1, S 2, …, S k+2 ] = (½) k+2 t = t’ = SiSi

April 20, 2004CS151 Lecture 728 Unique solutions –Pr[t agrees with some t’ on S 1,…, S k+2 ] ≤ (|T|-1)(½) k+2 < ½ –Pr[t satisfies S 1, S 2, …, S k+2 ] = (½) k+2 –Pr[t unique satisfying assignment of φ k+2 ] > (½) k+3 –sum over at least 2 k different t  T (disjoint events); claim follows.

April 20, 2004CS151 Lecture 729 Randomized complexity classes model: probabilistic Turing Machine –deterministic TM with additional read-only tape containing “coin flips” BPP (Bounded-error Probabilistic Poly-time) –L  BPP if there is a p.p.t. TM M: x  L  Pr y [M(x,y) accepts] ≥ 2/3 x  L  Pr y [M(x,y) rejects] ≥ 2/3 –“p.p.t” = probabilistic polynomial time

April 20, 2004CS151 Lecture 730 Randomized complexity classes RP (Random Polynomial-time) –L  RP if there is a p.p.t. TM M: x  L  Pr y [M(x,y) accepts] ≥ ½ x  L  Pr y [M(x,y) rejects] = 1 coRP (complement of Random Polynomial-time) –L  coRP if there is a p.p.t. TM M: x  L  Pr y [M(x,y) accepts] = 1 x  L  Pr y [M(x,y) rejects] ≥ ½

April 20, 2004CS151 Lecture 731 Randomized complexity classes “1/2” in RP, coRP definition unimportant –can replace by 1/poly(n) “2/3” in BPP definition unimportant –can replace by ½ + 1/poly(n) Why? error reduction –we will see simple error reduction by repetition –more sophisticated error reduction later These classes may capture “efficiently computable” better than P.

April 20, 2004CS151 Lecture 732 Error reduction for RP given L and p.p.t TM M: x  L  Pr y [M(x,y) accepts] ≥ ε x  L  Pr y [M(x,y) rejects] = 1 new p.p.t TM M’: –simulate M k/ε times, each time with independent coin flips –accept if any simulation accepts –otherwise reject

April 20, 2004CS151 Lecture 733 Error reduction x  L  Pr y [M(x,y) accepts] ≥ ε x  L  Pr y [M(x,y) rejects] = 1 if x  L: –probability a given simulation “bad” ≤ (1 – ε) –probability all simulations “bad” ≤ (1–ε) (k/ε) ≤ e -k Pr y’ [M’(x, y’) accepts] ≥ 1 – e -k if x  L: Pr y’ [M’(x,y’) rejects] = 1

April 20, 2004CS151 Lecture 734 Error reduction for BPP given L, and p.p.t. TM M: x  L  Pr y [M(x,y) accepts] ≥ ½ + ε x  L  Pr y [M(x,y) rejects] ≥ ½ + ε new p.p.t. TM M’: –simulate M k/ε 2 times, each time with independent coin flips –accept if majority of simulations accept –otherwise reject

April 20, 2004CS151 Lecture 735 Error reduction for BPP –X i random variable indicating “correct” outcome in i-th simulation (out of m = k/ε 2 ) Pr[X i = 1] ≥ ½ + ε Pr[X i = 0] ≤ ½ - ε –E[X i ] ≥ ½+ε –X = Σ i X i –μ = E[X] ≥ (½ + ε)m –Chernoff: Pr[X ≤ m/2] ≤ 2 -  (ε 2 μ)

April 20, 2004CS151 Lecture 736 Error reduction for BPP x  L  Pr y [M(x,y) accepts] ≥ ½ + ε x  L  Pr y [M(x,y) rejects] ≥ ½ + ε –if x  L Pr y’ [M’(x, y’) accepts] ≥ 1 – (½)  (k) –if x  L Pr y’ [M’(x,y’) rejects] ≥ 1 – (½)  (k)

April 20, 2004CS151 Lecture 737 Randomized complexity classes One more important class: ZPP (Zero-error Probabilistic Poly-time) –ZPP = RP  coRP –Pr y [M(x,y) outputs “fail”] ≤ ½ –otherwise outputs correct answer

April 20, 2004CS151 Lecture 738 Randomized complexity classes We have shown: –polynomial identity testing is in coRP –a poly-time algorithm for detecting unique solutions to SAT implies NP = RP

April 20, 2004CS151 Lecture 739 Relationship to other classes all these classes contain P –they can simply ignore the tape with coin flips all are in PSPACE –can exhaustively try all strings y –count accepts/rejects; compute probability RP  NP (and coRP  coNP) –multitude of accepting computations –NP requires only one

April 20, 2004CS151 Lecture 740 Relationship to other classes P RPcoRP NPcoNP PSPACE BPP

April 20, 2004CS151 Lecture 741 BPP How powerful is BPP? We have seen an example of a problem in BPP that we only know how to solve in EXP. Is randomness a panacea for intractability?

April 20, 2004CS151 Lecture 742 BPP It is not known if BPP = EXP (or even NEXP!) –but there are strong hints that it does not Is there a deterministic simulation of BPP that does better than brute-force search? –yes, if allow non-uniformity Theorem (Adelman): BPP  P/poly

April 20, 2004CS151 Lecture 743 BPP and Boolean circuits Proof: –language L  BPP –error reduction gives TM M such that if x  L Pr y [M(x, y) accepts] ≥ 1 – (½) |x| 2 if x  L Pr y [M(x, y) rejects] ≥ 1 – (½) |x| 2

April 20, 2004CS151 Lecture 744 BPP and Boolean circuits –say “y is bad for x” if M(x,y) gives incorrect answer –for fixed x: Pr y [y is bad for x] ≤ (½) |x| 2 –Pr y [y is bad for some x] ≤ 2 |x| (½) |x| 2 < 1 –Conclude: there exists some y on which M(x, y) is always correct –build circuit for M, hardwire this y

April 20, 2004CS151 Lecture 745 BPP and Boolean circuits Does BPP = EXP ? Adelman’s Theorem shows: BPP = EXP implies EXP  P/poly If you believe that randomness is all-powerful, you must also believe that non-uniformity gives an exponential advantage.

April 20, 2004CS151 Lecture 746 BPP Next: further explore the relationship between randomness and nonuniformity Main tool: pseudo-random generators