Coin flipping from a cosmic source OR Error correction of truly random bits Elchanan MosselRyan O’Donnell Microsoft Research MIT (now at Berkeley)

Slides:



Advertisements
Similar presentations
Optimal Lower Bounds for 2-Query Locally Decodable Linear Codes Kenji Obata.
Advertisements

1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
The Average Case Complexity of Counting Distinct Elements David Woodruff IBM Almaden.
Ulams Game and Universal Communications Using Feedback Ofer Shayevitz June 2006.
Extracting Randomness From Few Independent Sources Boaz Barak, IAS Russell Impagliazzo, UCSD Avi Wigderson, IAS.
Subhash Khot IAS Elchanan Mossel UC Berkeley Guy Kindler DIMACS Ryan O’Donnell IAS.
Foundations of Cryptography Lecture 2: One-way functions are essential for identification. Amplification: from weak to strong one-way function Lecturer:
Many-to-one Trapdoor Functions and their Relations to Public-key Cryptosystems M. Bellare S. Halevi A. Saha S. Vadhan.
Individual Position Slides: Jonathan Katz (University of Maryland) (Apologies I can’t be here in person)
On the robustness of dictatorships: spectral methods. Ehud Friedgut, Hebrew University, Jerusalem.
Computational Applications of Noise Sensitivity Ryan O’Donnell.
Circuit and Communication Complexity. Karchmer – Wigderson Games Given The communication game G f : Alice getss.t. f(x)=1 Bob getss.t. f(y)=0 Goal: Find.
Inapproximability of MAX-CUT Khot,Kindler,Mossel and O ’ Donnell Moshe Ben Nehemia June 05.
The Communication Complexity of Approximate Set Packing and Covering
On the Density of a Graph and its Blowup Raphael Yuster Joint work with Asaf Shapira.
Foundations of Cryptography Lecture 10 Lecturer: Moni Naor.
Uniqueness of Optimal Mod 3 Circuits for Parity Frederic Green Amitabha Roy Frederic Green Amitabha Roy Clark University Akamai Clark University Akamai.
Copyright © Cengage Learning. All rights reserved.
Universal Communication Brendan Juba (MIT) With: Madhu Sudan (MIT)
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Eran Omri, Bar-Ilan University Joint work with Amos Beimel and Ilan Orlov, BGU Ilan Orlov…!??!!
Section 7.4: Closures of Relations Let R be a relation on a set A. We have talked about 6 properties that a relation on a set may or may not possess: reflexive,
Majority and Minority games. Let G be a graph with all degrees odd. Each vertex is initially randomly assigned a colour (black or white), and at each.
Chain Rules for Entropy
Dictator tests and Hardness of approximating Max-Cut-Gain Ryan O’Donnell Carnegie Mellon (includes joint work with Subhash Khot of Georgia Tech)
The number of edge-disjoint transitive triples in a tournament.
Visual Recognition Tutorial
INFINITE SEQUENCES AND SERIES
1 Optimization problems such as MAXSAT, MIN NODE COVER, MAX INDEPENDENT SET, MAX CLIQUE, MIN SET COVER, TSP, KNAPSACK, BINPACKING do not have a polynomial.
The 1’st annual (?) workshop. 2 Communication under Channel Uncertainty: Oblivious channels Michael Langberg California Institute of Technology.
Avraham Ben-Aroya (Tel Aviv University) Oded Regev (Tel Aviv University) Ronald de Wolf (CWI, Amsterdam) A Hypercontractive Inequality for Matrix-Valued.
13. The Weak Law and the Strong Law of Large Numbers
CMSC 414 Computer and Network Security Lecture 6 Jonathan Katz.
Lecture 20: April 12 Introduction to Randomized Algorithms and the Probabilistic Method.
On Everlasting Security in the Hybrid Bounded Storage Model Danny Harnik Moni Naor.
Foundations of Cryptography Lecture 2 Lecturer: Moni Naor.
Introduction to AEP In information theory, the asymptotic equipartition property (AEP) is the analog of the law of large numbers. This law states that.
The importance of sequences and infinite series in calculus stems from Newton’s idea of representing functions as sums of infinite series.  For instance,
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology The Weak Law and the Strong.
1 2. Independence and Bernoulli Trials Independence: Events A and B are independent if It is easy to show that A, B independent implies are all independent.
Edge-disjoint induced subgraphs with given minimum degree Raphael Yuster 2012.
Channel Capacity.
The Integers. The Division Algorithms A high-school question: Compute 58/17. We can write 58 as 58 = 3 (17) + 7 This forms illustrates the answer: “3.
The Selection Problem. 2 Median and Order Statistics In this section, we will study algorithms for finding the i th smallest element in a set of n elements.
Great Theoretical Ideas in Computer Science.
MultiModality Registration Using Hilbert-Schmidt Estimators By: Srinivas Peddi Computer Integrated Surgery II April 27 th, 2001 Final Presentation.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
In section 11.9, we were able to find power series representations for a certain restricted class of functions. Here, we investigate more general problems.
1 Rainbow Decompositions Raphael Yuster University of Haifa Proc. Amer. Math. Soc. (2008), to appear.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
1 Decomposition into bipartite graphs with minimum degree 1. Raphael Yuster.
12 INFINITE SEQUENCES AND SERIES. In general, it is difficult to find the exact sum of a series.  We were able to accomplish this for geometric series.
Hedonic Clustering Games Moran Feldman Joint work with: Seffi Naor and Liane Lewin-Eytan.
Secret Sharing Non-Shannon Information Inequalities Presented in: Theory of Cryptography Conference (TCC) 2009 Published in: IEEE Transactions on Information.
MA/CSSE 473 Day 10 Primality Testing. MA/CSSE 473 Day 10 In-class exam: Friday, Sept 28 –You may bring a two-sided 8.5x11 inch piece of paper containing.
Approximation Algorithms based on linear programming.
Sorting by placement and Shift Sergi Elizalde Peter Winkler By 資工四 B 周于荃.
Theory of Computational Complexity M1 Takao Inoshita Iwama & Ito Lab Graduate School of Informatics, Kyoto University.
Random Access Codes and a Hypercontractive Inequality for
Probabilistic Algorithms
Introduction to Randomized Algorithms and the Probabilistic Method
Sensitivity of voting coin tossing protocols, Nov 1
Sensitivity of voting schemes and coin tossing protocols
Linear sketching with parities
When are Fuzzy Extractors Possible?
The Curve Merger (Dvir & Widgerson, 2008)
Linear sketching over
Summarizing Data by Statistics
When are Fuzzy Extractors Possible?
Linear sketching with parities
Presentation transcript:

Coin flipping from a cosmic source OR Error correction of truly random bits Elchanan MosselRyan O’Donnell Microsoft Research MIT (now at Berkeley)

A new problem We consider a new problem motivated by ideas in cryptography, coding theory, collective coin flipping, and noise sensitivity. We prove some results using probability, convexity, Fourier analysis, and discrete symmetrization. Many open problems remain.

The problem Alice Bob Cindy Kate x (n bits) y y y ° ° ° y k o o o first bit 0

Broadcast with ε errors Alice Bob Cindy Kate x (n bits) y y y ° ° ° y k o o o first bit

1 Broadcast with ε errors Alice Bob Cindy Kate x (n bits) y y y ° ° ° y k o o o majority

The parameters n bit uniform random “source” string x k parties who cannot communicate, but wish to agree on a uniformly random bit ε each party gets an independently corrupted version y i, each bit flipped independently with probability ε f (or f 1 … f k ): balanced “protocol” functions

Our goal For each n, k, ε, find the best protocol function f (or functions f 1 …f k ) which maximize the probability that all parties agree on the same bit.

Notation We’re interested in the probability (over choice of x and broadcast corruptions) that all parties agree. We write: P (f 1, …, f k ; ε) = Pr[f 1 (y 1 ) = ··· = f k (y k )], P k (f; ε)in the case f = f 1 = ··· = f k.

Motivation Original motivation: The “Everlasting Security” cryptographic protocol of Ding and Rabin [DR01]. In this model, many players want shared access to a random string. Requires a satellite or other cosmic source to broadcast trillions (!) of random bits per second. Errors in reception seem quite likely.

Motivation Natural question for the problem of error- correction in a broadcast channel. Of course, when the source is truly random, error correction is impossible. However we don’t require that all parties recover the original info with high probability, only that they attain some shared info with high probability and this mutual info has high entropy.

Motivation Similar to non-cryptographic collective coin- flipping problems [BL90,…, Dod00]. In these, a number of players want to agree on a random coin toss. However some players are malicious and corrupt bits arbitrarily. Two difference: 1. We assume random corruptions, not adversarial. 2. Our players cannot communicate.

Motivation Finally, the problem is intimately related to the study of noise sensitivity of boolean functions [KKL88, Hås97, BKS98, BJT99, Bou01, KS03, O02, MO02, KOS02, BMOS03,…]: this is the study of Pr[f(x) = f(y 1 )]. Technical aside: Noise sensitivity is essentially given by ||T ε (f)|| 2, where T ε is the linear operator from the Bonami- Beckner inequality. Our problem is essentially the study of ||T ε (f)|| k.

Intuition Suppose all players use the same balanced function f. In some sense, we want f to be the least noise sensitive balanced function possible. Normally, this is the first-bit dictator function. But if there are many players, we’d rather have a function which has a few points which are extremely noise-stable, rather than having all points fairly noise-stable…

Intuition – cont’d When f(x) = x 1, every source string is equally good; for each player, the probability its first bit doesn’t flip is 1-ε so the probability of success is something like (1-ε) k. When f(x) = majority, there are a few source strings, like 1111· · ·1, which are extremely good. So although majority is more noise sensitive “on the average,” it can be better in our problem if k is large.

Things harder than they seem? One theme we will allude to throughout the talk is that certain elements of this problem were more difficult or more counterintuitive than Elchanan and I expected – Some things we thought were obvious required or seemed to require nontrivial proofs; some things we thought were obvious weren’t even true!

About protocols For example, recall that we want the parties’ bits, when agreed upon, to be uniformly random. To get this, we restricted protocol functions to balanced. However this is neither necessary nor sufficient! In particular, for n = 5 and k = 3, there is a balanced function f such that, if all players use f, they are more likely to agree on 1 than on 0!

Antisymmetric protocols To get agreed-upon bits to be uniform, it suffices for functions be antisymmetric: f i ( x ) = f i (x). Proof: Pr[f 1 (y 1 ) = ··· = f k (y k ) = 1] = Pr[f 1 (y 1 ) = ··· = f k (y k ) = 0] =Pr[f 1 (y 1 ) = ··· = f k (y k ) = 0]. So we can study antisymmetric protocols instead if we like, but often studying merely balanced protocols is okay too.

Our results We first show that all players should use the same function, and it should have certain monotonicity properties. When k = 2 or 3, the first-bit function is best. For fixed n, when k→ ∞ majority is best, and when ε→0 and ε→½, the first-bit is best. For unbounded n, things get harder… in general we don’t know the best function, but we can give a lower bound for P k (f; ε).

Players should use same fcn. First, as expected, all parties should use the same function: Theorem 1: Fix n, k, ε and also a class of functions C for the parties’ functions to come from. Then every protocol which maximizes P (f 1, …, f k ; ε) has f 1 = ··· = f k. Proof: Convexity.

One page proof sketch Let C = {g 1, …, g m }, and suppose t i parties use g i, for i=1…m. We have that the t i ’s are integers and also: t i ≥ 0 and t 1 + ··· + t m = k. ( * ) The success probability which we want to maximize is a convex function of the t i ’s. Hence its maximum occurs at a vertex of ( * ), which is a point (0, …, 0, k, 0, …, 0), which is already integral.

For k=2,3, f(x) = x 1 is best Theorem 2: For k = 2, 3 and for all n, ε, the unique best protocol is for the parties to use f(x) = x 1. Proof: Fourier analysis. Comments: 1. If the players can be assumed to use the same function, the k=2 case is folklore. 2. By “unique,” we shall mean up to trivial reordering of indices and switching 0 and 1.

More on k=2, 3 Corollary: No error correction is possible for k=2, 3. Corollary: For all k, if the parties wish to maximize the expected number of agreements or the expected number of parties in the majority, they should all use f(x) = x 1. Proof: E[# (i,j) : f(y i ) = f(y j )] = ( ) Pr[f(y i ) = f(y j )]. n2n2

One page proof sketch for k = 2 When k = 2, we can think of party 1 as having the “true” random bits and party 2 as having an ε'-corruption. Thus the success probability is just the noise stability of f. For f balanced, this is: αΣ |S|≥1 (1-2ε') |S| f(S) 2, so best function has Fourier weight all on level 1. The k = 3 case reduces to k = 2 by a trick.

Any maximizing f has a special form: Theorem 3: For all k, n, ε, any f maximizing P k (f; ε) is left-monotone. Proof: Steiner symmetrization (shifting). Remark: This is again up to trivial permutations and switching 0 and 1. A left-monotone function is one satisfying f(x1y) ≥ f(x0y) and f(x10y) ≥ f(x01y) x,y. Properties of the best function A

Fixed ε, n; k→ ∞ For k > 3, you can just do better than f(x) = x 1 : Theorem 4: For all fixed ε and n (odd), for all sufficiently large k, the unique best protocol is f = MAJ n. Proof: Elementary probability and coupling. Remark: In this case, the probability of success = Θ( (1 – Pr[Bin(n,ε) > n/2]) k ), as compared to Θ( (1 – ε) k ) for f = x 1.

One page proof sketch intuitively, if n is fixed and k is very large, in most cases it’s extremely unlikely all agree to have a chance of success, must get a very helpful source string success probability indeed controlled by the success probability for the best source x since f can be assumed monotone, the best source string is the all 1’s string in this case, the best function is clearly MAJ n.

Fixed n, k; ε → 0, ½. Theorem 4 was for fixed n, ε and k → ∞. Dually: Theorem 5: Fix n and k. Then for ε sufficiently close to 0 and for ε sufficiently close to ½, the unique best protocol is f = x 1. Proof: Isoperimetry for ε near 0, Fourier analysis for ε near ½.

One page proof sketch for ε → 0 When ε is extremely tiny, it’s almost as though there is just a single corruption error among all y 1, …, y k. In this case, we just want to maximize the probability that this one corruption doesn’t change the value of f. This is equivalent to minimizing f’s “edge boundary.” By an isoperimetric theorem, the best f is the cube, f(x) = x 1.

Unbounded n As for k, ε fixed and n→ ∞, this is the heart of the problem and it seems quite difficult. Here we tend to imagine ε fixed and k→ ∞, but n is allowed to be unbounded in terms of k. It seemed to us from Theorem 4 that in this case, the probability of success should go to 0 exponentially quickly as k→ ∞. But…!

Polynomial decay We were unable to prove this because, in fact, the decay is at worst polynomial: Theorem 6: Fix ε. Then there is a sequence (n k ) such that: P k (MAJ n k ; ε) ≥ Ω(k -2/(1-2ε) ² ). Proof: Use normal approximation. Shameful fact: We still believe that the success probability must go to 0 as k→ ∞ but we can’t prove it! ~

Using majorities So far, in all of our theorems either MAJ 1 or MAJ n has been the best function. Unfortunately, it’s not true that one of these is always best. Theorem 7: There exist particular k and ε such that neither MAJ 1 nor MAJ n is the best majority function protocol. Indeed, P k (MAJ r ; ε) is not even unimodular in r! Proof: Computer-assisted.

Are majorities best? Still, in every case we know and every case considered by computer, some majority function has been best. Is this always the case? We present the two opposing conjectures on this intriguing question: Conjecture M: For a particular k, ε, and odd n, there is an antisymmetric function strictly better than all majority functions. Conjecture O: The best antisymmetric [balanced?] function is always a majority.

Wrap-up In conclusion, we think the “cosmic coin flipping” problem is a nice one to think about, and one that presents many intriguing open problems. We believe that some may be easy to resolve, whereas some might require much more heavy-duty techniques; perhaps some deeper isoperimetry ideas or the Bonami- Beckner inequality.

Open problems 1.Show that for fixed ε, when k→ ∞ and n is allowed to be unbounded, the success probability goes to 0. 2.Show that for all k, ε, as n→ ∞, the best majority is MAJ n, up to a universal constant. 3.Show that MAJ 1 is best for k ≤ … 9? 4.Show that the best weighted threshold function is always a majority. 5.Prove Conjecture M or Conjecture O.