On Gallager’s problem: New Bounds for Noisy Communication. Navin Goyal & Mike Saks Joint work with Guy Kindler Microsoft Research.

Slides:



Advertisements
Similar presentations
1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Advertisements

Optimal Space Lower Bounds for All Frequency Moments David Woodruff MIT
The Polynomial Method In Quantum and Classical Computing Scott Aaronson (MIT) OPEN PROBLEM.
Lower Bounds for Local Search by Quantum Arguments Scott Aaronson (UC Berkeley) August 14, 2003.
How to Solve Longstanding Open Problems In Quantum Computing Using Only Fourier Analysis Scott Aaronson (MIT) For those who hate quantum: The open problems.
Routing Complexity of Faulty Networks Omer Angel Itai Benjamini Eran Ofek Udi Wieder The Weizmann Institute of Science.
A threshold of ln(n) for approximating set cover By Uriel Feige Lecturer: Ariel Procaccia.
The Communication Complexity of Approximate Set Packing and Covering
Foundations of Cryptography Lecture 10 Lecturer: Moni Naor.
Nearest Neighbor Search in High Dimensions Seminar in Algorithms and Geometry Mica Arie-Nachimson and Daniel Glasner April 2009.
Locally Decodable Codes from Nice Subsets of Finite Fields and Prime Factors of Mersenne Numbers Kiran Kedlaya Sergey Yekhanin MIT Microsoft Research.
Gillat Kol (IAS) joint work with Ran Raz (Weizmann + IAS) Interactive Channel Capacity.
Introduction to PCP and Hardness of Approximation Dana Moshkovitz Princeton University and The Institute for Advanced Study 1.
Visual Recognition Tutorial
CS151 Complexity Theory Lecture 6 April 15, 2015.
On the tightness of Buhrman- Cleve-Wigderson simulation Shengyu Zhang The Chinese University of Hong Kong On the relation between decision tree complexity.
CS151 Complexity Theory Lecture 7 April 20, 2004.
Analysis of Boolean Functions Fourier Analysis, Projections, Influence, Junta, Etc… And (some) applications Slides prepared with help of Ricky Rosen.
Putting a Junta to the Test Joint work with Eldar Fischer, Dana Ron, Shmuel Safra, and Alex Samorodnitsky Guy Kindler.
Putting a Junta to the Test Joint work with Eldar Fischer & Guy Kindler.
Fourier Analysis, Projections, Influences, Juntas, Etc…
The 1’st annual (?) workshop. 2 Communication under Channel Uncertainty: Oblivious channels Michael Langberg California Institute of Technology.
Randomized and Quantum Protocols in Distributed Computation Michael Ben-Or The Hebrew University Michael Rabin’s Birthday Celebration.
Message Passing for the Coloring Problem: Gallager Meets Alon and Kahale Sonny Ben-Shimon and Dan Vilenchik Tel Aviv University AofA June, 2007 TexPoint.
Correlation Immune Functions and Learning Lisa Hellerstein Polytechnic Institute of NYU Brooklyn, NY Includes joint work with Bernard Rosell (AT&T), Eric.
Fourier Analysis of Boolean Functions Juntas, Projections, Influences Etc.
DAST 2005 Week 4 – Some Helpful Material Randomized Quick Sort & Lower bound & General remarks…
Foundations of Privacy Lecture 11 Lecturer: Moni Naor.
Collecting Correlated Information from a Sensor Network Micah Adler University of Massachusetts, Amherst.
Noise, Information Theory, and Entropy
Foundations of Cryptography Lecture 2 Lecturer: Moni Naor.
Quantum Computing MAS 725 Hartmut Klauck NTU TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A.
Streaming Algorithms Piotr Indyk MIT. Data Streams A data stream is a sequence of data that is too large to be stored in available memory Examples: –Network.
Zeev Dvir Weizmann Institute of Science Amir Shpilka Technion Locally decodable codes with 2 queries and polynomial identity testing for depth 3 circuits.
Many random walks are faster than one Noga AlonTel Aviv University Chen AvinBen Gurion University Michal KouckyCzech Academy of Sciences Gady KozmaWeizmann.
One-way multi-party communication lower bound for pointer jumping with applications Emanuele Viola & Avi Wigderson Columbia University IAS work done while.
1 Private codes or Succinct random codes that are (almost) perfect Michael Langberg California Institute of Technology.
Asymmetric Communication Complexity And its implications on Cell Probe Complexity Slides by Elad Verbin Based on a paper of Peter Bro Miltersen, Noam Nisan,
Fall 2013 CMU CS Computational Complexity Lectures 8-9 Randomness, communication, complexity of unique solutions These slides are mostly a resequencing.
Umans Complexity Theory Lectures Lecture 7b: Randomization in Communication Complexity.
CS 3343: Analysis of Algorithms Lecture 25: P and NP Some slides courtesy of Carola Wenk.
1 CSC 421: Algorithm Design & Analysis Spring 2014 Complexity & lower bounds  brute force  decision trees  adversary arguments  problem reduction.
Forrelation: A Problem that Optimally Separates Quantum from Classical Computing.
Communication Complexity Guy Feigenblat Based on lecture by Dr. Ely Porat Some slides where adapted from various sources Complexity course Computer science.
Smooth Boolean Functions are Easy: Efficient Algorithms for Low-Sensitivity Functions Rocco Servedio Joint work with Parikshit Gopalan (MSR) Noam Nisan.
1 On the Channel Capacity of Wireless Fading Channels C. D. Charalambous and S. Z. Denic School of Information Technology and Engineering, University of.
Jonathan Katz University of Maryland Andrew Lindell Aladdin Knowledge Systems and Bar-Ilan University 04/08/08 CRYP-108 Aggregate Message- Authentication.
Gillat Kol (IAS) joint work with Anat Ganor (Weizmann) Ran Raz (Weizmann + IAS) Exponential Separation of Information and Communication.
Why almost all satisfiable k - CNF formulas are easy? Danny Vilenchik Joint work with A. Coja-Oghlan and M. Krivelevich.
Error-Correcting Code
Space Complexity. Reminder: P, NP classes P is the class of problems that can be solved with algorithms that runs in polynomial time NP is the class of.
The Message Passing Communication Model David Woodruff IBM Almaden.
Analysis of Boolean Functions and Complexity Theory Economics Combinatorics …
Information Complexity Lower Bounds
New Characterizations in Turnstile Streams with Applications
Worst case to Average case Reductions for Polynomials
Lecture 22: Linearity Testing Sparse Fourier Transform
Effcient quantum protocols for XOR functions
Lecture 10: Sketching S3: Nearest Neighbor Search
Tight Fourier Tails for AC0 Circuits
General Strong Polarization
Topic 7: Pseudorandom Functions and CPA-Security
Linear sketching with parities
The Curve Merger (Dvir & Widgerson, 2008)
Linear sketching over
General Strong Polarization
Linear sketching with parities
Chapter 11 Limitations of Algorithm Power
CS21 Decidability and Tractability
Switching Lemmas and Proof Complexity
Presentation transcript:

On Gallager’s problem: New Bounds for Noisy Communication. Navin Goyal & Mike Saks Joint work with Guy Kindler Microsoft Research

Ambrose Bierce 1842 – 1914(?) “Noise is t he chief product and the authenticating sign of civilization”  In CS: Noise appears in the study of information theory, network design, learning theory, cryptography, quantum computation, hardness of approximation, theory of social choice, embeddings of metric spaces, privacy in databases…

In this talk  [El Gamal ’84]: The noisy broadcast network model.  [Gallager ’88]: n ¢ loglog(n) algorithm for identity.  Main result: Gallager’s algorithm is tight.  Proof by reduction: Generalized noisy decision trees (gnd-trees). Lower bound for gnd-trees.

First, a Fourier-analytic result  Definition (Fourier): Let f:{-1,1} n ! {-1,1} be a Boolean function. The i’th Fourier coefficient of f: f i = E x » U [f(x) ¢ x i ].  [Talagrand ’96]: Let p= Pr x » U [f(x)=1], (p<1/2). Then  i (f i ) 2 · p 2 log(1/p).  Crucial for our result! (as hinted in slide #26..)

What next: What next: Communication under noise - examples The noisy broadcast model Gallager: the algorithm and the problem Gnd-trees: Generalized Noisy Decision Trees Our results About the proof

01100 Noisy computation: case 1 1.Noiseless channel: n transmissions. 2.naïve: n ¢ log(n) (error is polynomially small in n) 3.[Shannon ’48]: c ¢ n (error is exponentially small in n) Aggregation of bits: Big advantage

y=10101 Noisy computation: case 2 x=01100 Goal: compute f(x,y) 1.Noiseless channel: k transmissions. 2. naïve: k ¢ log(k) 3.[Schulman ’96]: c ¢ k (error is exponentially small in k)

The Noisy Broadcast Model [El Gamal ’84] x1x1x1x1 x1x1x1x1 x5x5x5x5 x5x5x5x5 x4x4x4x4 x4x4x4x4 x3x3x3x3 x3x3x3x3 x2x2x2x2 x2x2x2x2 x6x6x6x6 x6x6x6x6 x7x7x7x7 x7x7x7x7 x8x8x8x8 x8x8x8x8 x9x9x9x9 x9x9x9x9 x 10  Input: x 1,..,x n.  One bit transmitted at a time.  Error rate:  (small const.).  Goal: compute g(x 1,..,x n ).  In this talk: we want to compute x 1,..,x n. Order of transmissions is predefined

Some history  Computing identity: Naïve solution: n log n (repetition) [Gallager ’88]: n loglog n.  [Yao 97]: Try thresholds first.  [KM ’98]: Any threshold in O(n).  In adversarial model: [FK ’00]: OR in O(n ¢ log * n). [N ’04]: OR in O(n). x1x1x1x1 x1x1x1x x5x5x5x5 x5x5x5x5 x4x4x4x4 x4x4x4x4 x3x3x3x3 x3x3x3x3 x2x2x2x2 x2x2x2x2 x6x6x6x6 x6x6x6x6 x7x7x7x7 x7x7x7x7 x8x8x8x8 x8x8x8x8 x9x9x9x9 x9x9x9x9 x1x1x1x1 x1x1x1x1 Fails for “adversarial noise”. Gallager’s problem: Can this be made linear?

what’s next: what’s next: Communication under noise - examples The noisy broadcast model Gallager: an algorithm and a problem Gnd-trees: Generalized Noisy Decision Trees Statement of results About the proof

g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 ) Generalized Noisy Decision (gnd) Trees  Input: x, but access is to noisy copies x 1,x 2, x 3 … x i =x © N i (N i flips x j w.p.  ) x i =x © N i (N i flips x j w.p.  )  Any Boolean queries! =“01” v =“01” f v : Boolean function Goal: compute g(x), Goal: compute g(x), minimizing depth(T) minimizing depth(T)

Generalized Noisy Decision (gnd) Trees  Noisy decision trees [FPRU ‘94]: Query noisy coordinates of x. Query noisy coordinates of x.  Identity computable in nlog(n). g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 )

[FPRU] O(n) [FPRU] Some bounds for noisy trees functionnoisy trees OR [FPRU]  n) [FPRU] PARITY [FPRU]  n log n) [FPRU] MAJORITY [FPRU]  n log n) [FPRU]  n) [GKS] [KM * ]  n) [KM * ] IDENTITY [FPRU]  n log n) [FPRU]  n log n) [GKS] gnd-trees g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 ) (n)(n) (n)(n)

Our results  Main theorem: bound for identity in n.b. network.  Main theorem:   (n ¢ loglog(n)) bound for identity in n.b. network.  Lower-bound for gnd-tree :   nlog(n) Lower bound for computing identity in generalized decision trees.  Reduction theorem: kn time protocol in  -noise n.b. network ) 2kn depth gnd-tree for noise  ck.  Proof of main theorem: 2kn ¸  ck n log n 2k(1/  ) ck ¸ log n k=  ( loglog(n) )

what’s next: what’s next: About communication under noise The noisy broadcast model Gallager: the algorithm and the problem Generalized Noisy Decision Trees (gnd-trees) Our results About the proof

About the proof: About the proof: The reduction:  A series of transformations from a broadcast protocol into a gnd-tree protocol.

About the proof: About the proof: The reduction:  A series of transformations from a broadcast protocol into a gnd-tree protocol. Gnd-tree lower bound:  Defining a knowledge measure.  Bounding knowledge measure by depth of tree.

Lower bound for gnd-trees  Our claim: A gnd-tree which computes identity on x=x 1,..,x n requires  (   n ¢ log n) depth.  We actually prove: If depth(T) ·   n ¢ log n then Pr x » U [T returns x] <  (  ), (lim  ! 0  (  )=0)

The big picture  We prove: If depth(T) ·   n ¢ log n then Pr x » U [T returns x] <  (  ), ( lim  ! 0  (  )=0 )  Structure of such proofs: 1.Define: Knowledge measure M x (v) 2.Show: T correct only if w.h.p. M x (  ) > t 3.Show: If depth(T)<<nlog n, then w.h.p. M x (  )<t In our case: t = log(n), and typically M x (v,a) - M x (v) · 1/(  3 ¢ n)  is the leaf reached by T. g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 ) Disclaimer: We consider case where each noisy copy is queried once… (more work needed in general case) Disclaimer: We consider case where each noisy copy is queried once… (more work needed in general case)

Perceived probability  Perceived probability (“likelihood”) of x: L x (v)=Pr[x|visit(v)]  Pr[x|visit(v)] is “multiplicative”. g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 )

Knowledge measure: 1 st attempt  Log likelihood of x: LL x (v) = n + log(L x (v)) LL x (root) =0, LL x ()n – const LL x (  ) ¸ n – const  We’d like to show: Typically, LL x (v,a)-LL x (v) < 1/log(n).  But : After n coordinate queries, LL x ¼  (n).  Reason: x is quickly separated from far away points. Separating x from neighbors is the hardest. g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 )

Knowledge measure: seriously  Log likelihood “gradient” at x: M i x (v)= log(L x (v)) - log(L x © i (v)) M x (v)= AVG i (M i x (v)) = log(L x (v)) - AVG i ( log(L x © i (v)) ) = log(L x (v)) - AVG i ( log(L x © i (v)) ) M x (root)=0, M x () ¸ log(n) - c M x (root)=0, M x (  ) ¸ log(n) - c g(x)=y g(x)=.. f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 1 ) f  (x 1 ) f  (x 2 ) f  (x 2 ) All that is left: typical gain in M x is at most 1/n.

a =1. v f (x 5 ) v,1 v,0 v,1 Gain in knowledge measure

M i x (v,a)- M i x (v)= v f (x 5 ) v,0 v,1 a =1. log(L x (v,a)) - log(L x © i (v,a)) - ( log(L x (v)) - log(L x © i (v)) ) log(L x (v,a)) - log(L x © i (v,a)) - ( log(L x (v)) - log(L x © i (v)) )

Gain in knowledge measure M i x (v,a)- M i x (v) M x (v,a)-M x (v)=  The coup des grâce: For every query f v, x, E[M x (v,a)-M x (v)] · 1/(  3 n) E[(M x (v,a)-M x (v)) 2 ] · 1/(  3 n)  Proof: Adaptation of [Talagrand ‘96]. v f (x 5 ) v,0 v,1 Expression depends only on f, x !

Main open problem  Show lower bound for computing a Boolean function.  Not known even for a random function!  Generalize for other network designs.

Thank You !

Gallager’s solution, simplified 1.Partition to groups of size log(n) 2.Each player sends its bit loglog(n) times ,1, ,0,0

1 Gallager’s solution, simplified 1.Partition to groups of size log(n) 2.Each player sends its bit loglog(n) times ,0,

Gallager’s solution, simplified 1.Partition to groups of size log(n) 2.Each player sends its bit loglog(n) times.  W.h.p., in all groups, almost all players know all bits

Gallager’s solution, simplified 3.Each group transmits error correcting code of its bits: * Each player transmits a constant number of bits. 4.W.h.p. all players now know all bits of all groups ,1,1 0,1,10,1,0 1,0,0 suppose code(1001)=

The reduction  The program: Start with a noisy broadcast protocol with kn steps. Gradually, simlulate protocol in more “tree-like” models.  W.l.o.g., assume each node performs 10k transmissions.  first step: each transmission is replaced by three, only one of which is noisy.

The reduction  First step: each transmission is replaced by three, only one of which is noisy. Function of x 3, and of past receptions. b x3x3x3x3 x3x3x3x3 x 3, b(0),b(1) x3x3x3x3 x3x3x3x3 b(0), b(1) transmitted noise free.

The reduction  Second step: noisy transmissions moved to beginning of protocol. Function of x 3, and of past receptions. b x3x3x3x3 x3x3x3x3 b(0),b(1) x3x3x3x3 x3x3x3x3 x 3, x 3, x 3,.. x3x3x3x3 x3x3x3x3

The reduction  Second step: noisy transmissions moved to beginning of protocol.  After noisy phase: each player has 10k noisy copies of each bit.  Equivalent to having an  k -noisy copy of x. b(0),b(1) x3x3x3x3 x3x3x3x3 x 3, x 3, x 3,.. x3x3x3x3 x3x3x3x3

The reduction  Third step: each player begins with an  k -noisy copy of x.  Each transmission depends on transmitter’s noisy copy, and past transmissions (and perhaps a random decision). b(0),b(1) x 3, x 3, x 3,.. x©N3x©N3 x©N3x©N3 Equivalent to a gnd tree!

Gain in progress measure M i x (v)= log(Pr[x|visit(v)])-log(Pr[x © i|visit(v)]) v f (x 5 ) v,0 v,1

Gain in progress measure M i x (v) M i x (v) M i x (v,a)- M i x (v)= v f (x 5 ) v,0 v,1 a = f (x 5 ) : a random variable Only depends on f !.