Download presentation
Presentation is loading. Please wait.
Published byAbner McKenzie Modified over 9 years ago
1
Eric Allender Rutgers University Circuit Complexity meets the Theory of Randomness SUNY Buffalo, November 11, 2010
2
Eric Allender: Circuit Complexity meets the Theory of Randomness < 2 >< 2 > Today’s Goal: To raise awareness of the tight connection between two fields: – Circuit Complexity – Kolmogorov Complexity (the theory of randomness) And to show that this is useful. More generally, to spread the news about some exciting developments in the field of derandomization.
3
Eric Allender: Circuit Complexity meets the Theory of Randomness < 3 >< 3 > Derandomization Probabilistic algorithm Regular input Random bits b 1,b 2,… Where do these random bits come from?
4
Eric Allender: Circuit Complexity meets the Theory of Randomness < 4 >< 4 > Derandomization Probabilistic algorithm Regular input Random bits b 1,b 2,… Random bits are: Expensive Low-quality Hard to find Possibly non-existent!
5
Eric Allender: Circuit Complexity meets the Theory of Randomness < 5 >< 5 > Derandomization Probabilistic algorithm Regular input PseudoRandom bits b 1,b 2,… seed Generator g
6
Eric Allender: Circuit Complexity meets the Theory of Randomness < 6 >< 6 > A Jewel of Derandomization [Impagliazzo, Wigderson, 1997]: If there is a problem computable in time 2 n that requires circuits of size 2 εn, then there is a “good” generator that takes seeds of length O(log n). Thus, for example, if your favorite NP- complete problem requires big circuits (as we expect), then any probabilistic algorithm can be simulated deterministically with modest slow-down. (Run the generator on all of the poly-many seeds, and take the majority vote.)
7
Eric Allender: Circuit Complexity meets the Theory of Randomness < 7 >< 7 > A Jewel of Derandomization [Impagliazzo, Wigderson, 1997]: If there is a problem computable in time 2 n that requires circuits of size 2 εn, then there is a “good” generator that takes seeds of length O(log n). Stated another way (and introducing some notation), a very plausible hypothesis implies P = BPP.
8
Eric Allender: Circuit Complexity meets the Theory of Randomness < 8 >< 8 > A Jewel of Derandomization [Impagliazzo, Wigderson, 1997]: If there is a problem computable in time 2 n that requires circuits of size 2 εn, then there is a “good” generator that takes seeds of length O(log n). The proof is deep, and incorporates clever algorithms, deep combinatorics, and innovations in coding theory. …and it provides some great tools that shed insight into the nature of randomness.
9
Eric Allender: Circuit Complexity meets the Theory of Randomness < 9 >< 9 > Randomness Which of these sequences did I obtain by flipping a coin? 0101010101010101010101010101010 0110100111111100111000100101100 1101010001011100111111011001010 1/π =0.3183098861837906715377675267450 Each of these sequences is equally likely, in the sense of probability theory – which does not provide us with a way to talk about “randomness”.
10
Eric Allender: Circuit Complexity meets the Theory of Randomness Kolmogorov Complexity C(x) = min{|d| : U(d) = x} – U is a “universal” Turing machine Important property – Invariance: The choice of the universal Turing machine U is unimportant (up to an additive constant). x is random if C(x) ≥ |x|. C A (x) = min{|d| : U A (d) = x} Let’s make a connection between Kolmogorov complexity and circuit complexity.
11
Eric Allender: Circuit Complexity meets the Theory of Randomness Circuit Complexity Let D be a circuit of AND and OR gates (with negations at the inputs). Size(D) = # of wires in D. Size(f) = min{Size(D) : D computes f} We may allow oracle gates for a set A, along with AND and OR gates. Size A (f) = min{Size(D) : D A computes f}
12
Eric Allender: Circuit Complexity meets the Theory of Randomness Oracle Gates An “oracle gate” for B in a circuit: B
13
Eric Allender: Circuit Complexity meets the Theory of Randomness K-complexity ≈ Circuit Complexity There are some obvious similarities in the definitions. C(x) = min{|d| : U(d) = x} Size(f) = min{Size(D) : D computes f}
14
Eric Allender: Circuit Complexity meets the Theory of Randomness K-complexity ≈ Circuit Complexity There are some obvious similarities in the definitions. What are some differences? A minor difference: Size gives a measure of the complexity of functions, C gives a measure of the complexity of strings. – Given any string x, let f x be the function whose truth table is the string of length 2 log|x|, padded out with 0’s, and define Size(x) to be Size(f x ).
15
Eric Allender: Circuit Complexity meets the Theory of Randomness K-complexity ≈ Circuit Complexity There are some obvious similarities in the definitions. What are some differences? A minor difference: Size gives a measure of the complexity of functions, C gives a measure of the complexity of strings. A more fundamental difference: – C(x) is not computable; Size(x) is.
16
Eric Allender: Circuit Complexity meets the Theory of Randomness Lightning Review: Computability Here’s all we’ll need to know about computability: – The halting problem, everyone’s favorite uncomputable problem: – H = {(i,x) : M i halts on input x} – Every r.e. problem is poly-time many-one reducible to H. – One such r.e. problem: {x : C(x) < |x|}. – There is no time-bounded many-reduction in other direction, but there is a (large) time t Turing reduction (and hence C is not computable).
17
Eric Allender: Circuit Complexity meets the Theory of Randomness K-complexity ≈ Circuit Complexity There are some obvious similarities in the definitions. What are some differences? A minor difference: Size gives a measure of the complexity of functions, C gives a measure of the complexity of strings. A more fundamental difference: – C(x) is not computable; Size(x) is. In fact, there is a fascinating history, regarding the complexity of the Size function.
18
Eric Allender: Circuit Complexity meets the Theory of Randomness MCSP MCSP = {(x,i) : Size(x) < i}. MCSP is in NP, but is not known to be NP- complete. Some history: NP-completeness was discovered independently by in the 70s by Cook (in North America) and Levin (in Russia). Levin delayed publishing his results, because he wanted to show that MCSP was NP- complete, thinking that this was the more important problem.
19
Eric Allender: Circuit Complexity meets the Theory of Randomness MCSP MCSP = {(x,i) : Size(x) < i}. Why was Levin so interested in MCSP? In the USSR in the 70’s, there was great interest in problems requiring “perebor”, or “brute-force search”. For various reasons, MCSP was a focal point of this interest. It was recognized at the time that this was “similar in flavor” to questions about resource- bounded Kolmogorov complexity, but there were no theorems along this line.
20
Eric Allender: Circuit Complexity meets the Theory of Randomness MCSP MCSP = {(x,i) : Size(x) < i}. 3 decades later, what do we know about MCSP? If it’s complete under the “usual” poly-time many-one reductions, then BPP = P. MCSP is not believed to be in P. We showed: – Factoring is in BPP MCSP. – Every cryptographically-secure one-way function can be inverted in P MCSP /poly.
21
Eric Allender: Circuit Complexity meets the Theory of Randomness So how can K-complexity and Circuit complexity be the same? C(x) ≈ Size H (x), where H is the halting problem. For one direction, let U(d) = x. We need a small circuit (with oracle gates for H) for f x, where f x (i) is the i-th bit of x. This is easy, since {(d,i,b) : U(d) outputs a string whose i-th bit is b} is computably-enumerable. For the other direction, let Size H (f x ) = m. No oracle gate has more than m wires coming into it. Given a description of D (size not much bigger than m) and the m-bit number giving the size of {y in H : |y| ≤ m}, U can simulate D H and produce f x.
22
Eric Allender: Circuit Complexity meets the Theory of Randomness So how can K-complexity and Circuit complexity be the same? C(x) ≈ Size H (x), where H is the halting problem. So there is a connection between C(x) and Size(x) … …but is it useful? First, let’s look at decidable versions of Kolmogorov complexity.
23
Eric Allender: Circuit Complexity meets the Theory of Randomness Time-Bounded Kolmogorov Complexity The usual definition: C t (x) = min{|d| : U(d) = x in time t(|x|)}. Problems with this definition – No invariance! If U and U’ are different universal Turing machines, C t U and C t U’ have no clear relationship. – (One can bound C t U by C t’ U’ for t’ slightly larger than t – but nothing can be done for t’=t.) No nice connection to circuit complexity!
24
Eric Allender: Circuit Complexity meets the Theory of Randomness Time-Bounded Kolmogorov Complexity Levin’s definition: Kt(x) = min{|d|+log t : U(d) = x in time t}. Invariance holds! If U and U’ are different universal Turing machines, Kt U (x) and Kt U’ (x) are within log |x| of each other. And, there’s a connection to Circuit Complexity: – Let A be complete for E = Dtime(2 O(n) ). Then Kt(x) ≈ Size A (x).
25
Eric Allender: Circuit Complexity meets the Theory of Randomness Time-Bounded Kolmogorov Complexity Levin’s definition: Kt(x|φ) = min{|d|+log t : U(d,φ)=x in time t}. Why log t? – This gives an optimal search order for NP search problems. – This algorithm finds satisfying assignments in poly time, if any algorithm does: – On input φ, search through assignments y in order of increasing Kt(y|φ).
26
Eric Allender: Circuit Complexity meets the Theory of Randomness Time-Bounded Kolmogorov Complexity Levin’s definition: Kt(x) = min{|d|+log t : U(d) = x in time t}. Why log t? – This gives an optimal search order for NP search problems. – Adding t instead of log t would give every string complexity ≥ |x|. …So let’s look for a sensible way to allow the run-time to be much smaller.
27
Eric Allender: Circuit Complexity meets the Theory of Randomness Revised Kolmogorov Complexity C(x) = min{|d| : for all i ≤ |x| + 1, U(d,i,b) = 1 iff b is the i-th bit of x} (where bit # i+1 of x is *). – This is identical to the original definition. Kt(x) = min{|d|+log t : for all i ≤ |x| + 1, U(d,i,b) = 1 iff b is the i-th bit of x, in time t}. – The new and old definitions are within O(log |x|) of each other. Define KT(x) = min{|d|+t : for all i ≤ |x| + 1, U(d,i,b) = 1 iff b is the i-th bit of x, in time t}.
28
Eric Allender: Circuit Complexity meets the Theory of Randomness Kolmogorov Complexity is Circuit Complexity C(x) ≈ Size H (x). Kt(x) ≈ Size E (x). KT(x) ≈ Size(x). Other measures of complexity can be captured in this way, too: – Branching Program Size ≈ KB(x) = min{|d|+2 s : for all I ≤ |x| + 1, U(d,i,b) = 1 iff b is the i-th bit of x, in space s}.
29
Eric Allender: Circuit Complexity meets the Theory of Randomness Kolmogorov Complexity is Circuit Complexity C(x) ≈ Size H (x). Kt(x) ≈ Size E (x). KT(x) ≈ Size(x). Other measures of complexity can be captured in this way, too: – Formula Size ≈ KF(x) = min{|d|+2 t : for all I ≤ |x| + 1, U(d,i,b) = 1 iff b is the i-th bit of x, in time t}, for an alternating Turing machine U.
30
Eric Allender: Circuit Complexity meets the Theory of Randomness …but is this interesting? The result that Factoring is in BPP MCSP was first proved by observing that, in P MCSP, one can accept a large set of strings having large KT complexity: R KT = {x : KT(x) ≥ |x|}. (Basic Idea): There is a pseudorandom generator based on factoring, such that factoring is in BPP T for any test T that distinguishes truly random strings from pseudorandom strings. R KT is just such a test T.
31
Eric Allender: Circuit Complexity meets the Theory of Randomness Given a hard f, g is good Probabilistic algorithm Regular input PseudoRandom bits b 1,b 2,… seed Generator g
32
Eric Allender: Circuit Complexity meets the Theory of Randomness This idea has many variants. Consider R KT, R Kt, and R C. R KT is in coNP, and not known to be coNP hard. R C is not hard for NP under poly-time many-one reductions, unless P=NP. – How about more powerful reductions? – Is there anything interesting that we could compute quickly if C were computable, that we can’t already compute quickly? – Proof uses PRGs, Interactive Proofs, and the fact that an element of R C of length n can be found in But R C is undecidable! Perhaps H is in P relative to R C ?
33
Eric Allender: Circuit Complexity meets the Theory of Randomness This idea has many variants. Consider R KT, R Kt, and R C. R KT is in coNP, and not known to be coNP hard. R C is not hard for NP under poly-time many-one reductions, unless P=NP. – How about more powerful reductions? We show: – PSPACE is in P relative to R C. – NEXP is in NP relative to R C. – Proof uses PRGs, Interactive Proofs, and the fact that an element of R C of length n can be found in poly time, relative to R C [BFNV]. But R C is undecidable! Perhaps H is in P relative to R C ?
34
Eric Allender: Circuit Complexity meets the Theory of Randomness Relationship between H and R C Perhaps H is in P relative to R C ? This is still open. It is known that there is a computable time bound t such that H is in DTime(t) relative to R C [Kummer]. …but the bound t depends on the choice of U in the definition of C(x). We also showed: H is in P/poly relative to R C. Thus, in some sense, there is an “efficient” reduction of H to R C.
35
Eric Allender: Circuit Complexity meets the Theory of Randomness This idea has many variants. Consider R KT, R Kt, and R C. What about R Kt ? R Kt, is not hard for NP under poly-time many- one reductions, unless E=NE. – How about more powerful reductions? – We show: EXP = NP(R Kt ). – …and R Kt is complete for EXP under P/poly reductions. – Open if R Kt is in P!
36
Eric Allender: Circuit Complexity meets the Theory of Randomness A Very Different Complete Set R Kt, is not hard for NP under poly-time many- one reductions, unless E=NE. And it’s provably not complete for EXP under poly-time many-one reductions. Yet it is complete under P/poly reductions and under NP-Turing reductions. First example of a “natural” complete set that’s complete only using more powerful reductions.
37
Eric Allender: Circuit Complexity meets the Theory of Randomness A Very Different Complete Set R Kt, is not hard for NP under poly-time many- one reductions, unless E=NE. And it’s provably not complete for EXP under poly-time many-one reductions. (Sketch): Assume that R Kt is complete under poly-time many-one reductions, and let A be a language over 1* in EXP that’s not in P. (Such sets A exist.) Let f reduce A to R Kt. Thus f(1 n ) is in R Kt iff 1 n is in A. But Kt(f(1 n )) < O(log n), and thus the only way it can be in R Kt is if f(1 n ) is short (O(log n) bits). Thus A is in P, contrary to assumption.
38
Eric Allender: Circuit Complexity meets the Theory of Randomness Playing the Game in PSPACE We can also define a space-bounded measure KS, such that Kt(x) < KS(x) < KT(x). RKS is complete for PSPACE under BPP reductions (and, in fact, even under “ZPP” reductions, which implies completeness under NP reductions and P/poly reductions). [This makes use of an even stronger form of the Impagliazzo-Wigderson generator, which relies on the “hard” function being “downward self-reducible and random self-reducible”.]
39
Eric Allender: Circuit Complexity meets the Theory of Randomness Playing the Game Beyond EXP For essentially any “large” DTIME, NTIME, or DSPACE complexity class C, one can show that the problem of computing (or approximating) the SIZE A function (where A is the standard complete set for C) is complete for C under P/poly reductions. However, we obtain completeness under uniform notions of reducibility (such as NP- or ZPP- reducibility) only inside NEXP. Is this inherent? (E.g., is the inclusion “NEXP is contained in NP(R C )” optimal?)
40
Eric Allender: Circuit Complexity meets the Theory of Randomness Some Speculation Is this inherent? (E.g., is the inclusion “NEXP is contained in NP(R C )” optimal?) If so, then this could lead to characterizations of complexity classes in terms of efficient reductions to the (undecidable) set R C. …which would sufficiently strange, that it would certainly have to be useful!
41
Eric Allender: Circuit Complexity meets the Theory of Randomness Some Speculation Is this inherent? (E.g., is the inclusion “NEXP is contained in NP(R C )” optimal?) If so, then this could lead to characterizations of complexity classes in terms of efficient reductions to the (undecidable) set R C. New development: evidence that complexity classes can be characterized in terms of reductions to R C (or R K ). We know: NEXP is in NP relative to R K (no matter which univ. TM we use to define K(x)). No class bigger than EXPSPACE has this property.
42
Eric Allender: Circuit Complexity meets the Theory of Randomness Things to Remember There has been impressive progress in derandomization, so that now most people in the field would conjecture that BPP = P. The tools of derandomization can be viewed through the lens of Kolmogorov complexity, because of the close connection that exists between circuit complexity and Kolmogorov complexity.
43
Eric Allender: Circuit Complexity meets the Theory of Randomness Things to Remember This has led to new insights in complexity theory, such as: – Clarification of the complexity of Levin’s Kt function. – Presentation of fundamentally different classes of complete sets for PSPACE and EXP. – Clarification of the complexity of MCSP. It has also led to new insights in Kolmogorov complexity (such as the efficient reduction of H to R C ).
44
Eric Allender: Circuit Complexity meets the Theory of Randomness A Sampling of Open Questions Is MCSP complete for NP under P/poly reductions? (The analogous sets in PSPACE, EXP, NEXP, etc. are complete under P/poly reductions.) Is MCSP “complete” in some sense for cryptography? (I.e., can one show that secure crypto is possible ↔ MCSP is not in P/poly? The → direction is known. How about the ← direction? Can one build a crypto scheme based on MCSP?)
45
Eric Allender: Circuit Complexity meets the Theory of Randomness A Sampling of Open Questions Can one prove (unconditionally) that R Kt is not in P? (I know of no reason why this should be hard to prove.) – Note in this regard, that EXP = ZPP iff there is a “dense” set in P that contains no strings of Kt complexity < √n. Can one show that R Kt is not complete for EXP under poly-time Turing reductions?
46
Eric Allender: Circuit Complexity meets the Theory of Randomness A Sampling of Open Questions Can one improve the inclusions: – PSPACE is in P relative to R C. – NEXP is in NP relative to R C. Can one show that H is not poly-time Turing reducible to R C ?
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.