Communication Amid Uncertainty

Slides:



Advertisements
Similar presentations
Estimating Distinct Elements, Optimally
Advertisements

1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Tight Bounds for Distributed Functional Monitoring David Woodruff IBM Almaden Qin Zhang Aarhus University MADALGO Based on a paper in STOC, 2012.
Tight Bounds for Distributed Functional Monitoring David Woodruff IBM Almaden Qin Zhang Aarhus University MADALGO.
Limitations of Quantum Advice and One-Way Communication Scott Aaronson UC Berkeley IAS Useful?
The Average Case Complexity of Counting Distinct Elements David Woodruff IBM Almaden.
Optimal Bounds for Johnson- Lindenstrauss Transforms and Streaming Problems with Sub- Constant Error T.S. Jayram David Woodruff IBM Almaden.
Xiaoming Sun Tsinghua University David Woodruff MIT
Tight Lower Bounds for the Distinct Elements Problem David Woodruff MIT Joint work with Piotr Indyk.
Of 29 12/02/2013Purdue: Uncertainty in Communication1 Communication amid Uncertainty Madhu Sudan Microsoft, Cambridge, USA Based on: -Universal Semantic.
Approximate List- Decoding and Hardness Amplification Valentine Kabanets (SFU) joint work with Russell Impagliazzo and Ragesh Jaiswal (UCSD)
Of 24 11/20/2012TIFR: Deterministic Communication Amid Uncertainty1 ( Deterministic ) Communication amid Uncertainty Madhu Sudan Microsoft, New England.
Of 30 10/31/2013Cornell: Uncertainty in Communication1 Communication amid Uncertainty Madhu Sudan Microsoft, Cambridge, USA Based on: -Universal Semantic.
Of 30 09/16/2013PACM: Uncertainty in Communication1 Communication amid Uncertainty Madhu Sudan Microsoft, Cambridge, USA Based on: -Universal Semantic.
Inapproximability of MAX-CUT Khot,Kindler,Mossel and O ’ Donnell Moshe Ben Nehemia June 05.
The Communication Complexity of Approximate Set Packing and Covering
Gillat Kol (IAS) joint work with Ran Raz (Weizmann + IAS) Interactive Channel Capacity.
Of 7 10/01/2013LIDS Lunch: Communication Amid Uncertainty1 Communication Amid Uncertainty Madhu Sudan Microsoft Research.
Universal Communication Brendan Juba (MIT) With: Madhu Sudan (MIT)
Of 30 09/04/2012ITW 2012: Uncertainty in Communication1 Communication amid Uncertainty Madhu Sudan Microsoft, Cambridge, USA Based on: Universal Semantic.
Eran Omri, Bar-Ilan University Joint work with Amos Beimel and Ilan Orlov, BGU Ilan Orlov…!??!!
Of 27 01/06/2015CMI: Uncertain Communication1 Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on Juba, S. (STOC 2008, ITCS 2011) Juba,
Of 12 03/22/2012CISS: Compression w. Uncertain Priors1 Compression under uncertain priors Madhu Sudan Microsoft, New England Based on joint works with:
Lattices for Distributed Source Coding - Reconstruction of a Linear function of Jointly Gaussian Sources -D. Krithivasan and S. Sandeep Pradhan - University.
Of 35 05/16/2012CTW: Communication and Computation1 Communication amid Uncertainty Madhu Sudan Microsoft, Cambridge, USA Based on: Universal Semantic Communication.
Of 19 June 15, 2015CUHK: Communication Amid Uncertainty1 Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on joint works with Brendan.
The Price of Uncertainty in Communication Brendan Juba (Washington U., St. Louis) with Mark Braverman (Princeton)
Umans Complexity Theory Lectures Lecture 7b: Randomization in Communication Complexity.
Of 22 10/07/2015UMass: Uncertain Communication1 Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on Juba, S. (STOC 2008, ITCS 2011)
Of 22 10/30/2015WUSTL: Uncertain Communication1 Communication Amid Uncertainty Madhu Sudan Harvard Based on Juba, S. (STOC 2008, ITCS 2011) Juba, S. (STOC.
Data Stream Algorithms Lower Bounds Graham Cormode
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Imperfectly Shared Randomness
Communication Amid Uncertainty
Information Complexity Lower Bounds
Distributed Compression For Still Images
Game Theory Just last week:
New Characterizations in Turnstile Streams with Applications
And now for something completely different!
Communication Amid Uncertainty
General Strong Polarization
Communication amid Uncertainty
Universal Semantic Communication
Branching Programs Part 3
Background: Lattices and the Learning-with-Errors problem
General Strong Polarization
CS 154, Lecture 6: Communication Complexity
Turnstile Streaming Algorithms Might as Well Be Linear Sketches
Communication Amid Uncertainty
Linear sketching with parities
When are Fuzzy Extractors Possible?
Locally Decodable Codes from Lifting
Alternating tree Automata and Parity games
Uncertain Compression
Linear sketching over
General Strong Polarization
When are Fuzzy Extractors Possible?
Linear sketching with parities
The Byzantine Secretary Problem
Imperfectly Shared Randomness
Universal Semantic Communication
Universal Semantic Communication
Universal Semantic Communication
Communication Amid Uncertainty
CS21 Decidability and Tractability
What should I talk about? Aspects of Human Communication
General Strong Polarization
General Strong Polarization
Error Correction Coding
Presentation transcript:

Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on joint works with Brendan Juba, Oded Goldreich, Adam Kalai, Sanjeev Khanna, Elad Haramaty, Jacob Leshno, Clement Canonne, Venkatesan Guruswami, Badih Ghazi, Pritish Kamath, Ilan Komargodski and Pravesh Kothari. July 28, 2015 CMU: Communication Amid Uncertainty

Context in Communication Sender + Receiver share (huuuge) context In human comm: Language, news, Social In computer comm: Protocols, Codes, Distributions Helps compress communication Perfectly shared ⇒ Can be abstracted away. Imperfectly shared ⇒ What is the cost? How to study? July 28, 2015 CMU: Communication Amid Uncertainty

Communication Complexity The model (with shared randomness) 𝑥 𝑦 𝑓: 𝑥,𝑦 ↦Σ 𝑅=$$$ Usually studied for lower bounds. This talk: CC as +ve model. Alice Bob 𝐶𝐶 𝑓 =# bits exchanged by best protocol 𝑓(𝑥,𝑦) w.p. 2/3 July 28, 2015 CMU: Communication Amid Uncertainty

Modelling Shared Context + Imperfection 1 Many possibilities. Ongoing effort. Alice+Bob may have estimates of 𝑥 and 𝑦. More generally: 𝑥,𝑦 correlated. Knowledge of 𝑓 – function Bob wants to compute may not be exactly known to Alice! Shared randomness Alice + Bob may not have identical copies. 3 2 July 28, 2015 CMU: Communication Amid Uncertainty

Part 1: Uncertain Compression July 28, 2015 CMU: Communication Amid Uncertainty

Specific Motivation: Dictionary 𝑀 1 = 𝑤 11 , 𝑤 12 , … 𝑀 2 = 𝑤 21 , 𝑤 22 , … 𝑀 3 = 𝑤 31 , 𝑤 32 , … 𝑀 4 = 𝑤 41 , 𝑤 42 , … … Dictionary: maps words to meaning Multiple words with same meaning Multiple meanings to same word How to decide what word to use (encoding)? How to decide what a word means (decoding)? Common answer: Context Really Dictionary specifies: Encoding: context × meaning → word Decoding: context × word → meaning Context implicit; encoding/decoding works even if context used not identical! July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Context? In general complex notion … What does sender know/believe What does receiver know/believe Modifies as conversation progresses. Our abstraction: Context = Probability distribution on potential “meanings”. Certainly part of what the context provides; and sufficient abstraction to highlight the problem. July 28, 2015 CMU: Communication Amid Uncertainty

The (Uncertain Compression) problem Wish to design encoding/decoding schemes (𝐸/𝐷) to be used as follows: Sender has distribution 𝑃 on 𝑀 = {1,2,…,𝑁} Receiver has distribution 𝑄 on 𝑀 = {1,2,…,𝑁} Sender gets 𝑋∈𝑀 Sends 𝐸(𝑃,𝑋) to receiver. Receiver receives 𝑌 = 𝐸(𝑃,𝑋) Decodes to 𝑋 =𝐷(𝑄,𝑌) Want: 𝑋= 𝑋 (provided 𝑃,𝑄 close), While minimizing 𝐸𝑥 𝑝 𝑋←𝑃 |𝐸(𝑃,𝑋)| July 28, 2015 CMU: Communication Amid Uncertainty

Closeness of distributions: 𝑃 is Δ-close to 𝑄 if for all 𝑋∈𝑀, 1 2 Δ ≤ 𝑃 𝑋 𝑄 𝑋 ≤ 2 Δ 𝑃 Δ-close to 𝑄 ⇒ 𝐷(𝑃||𝑄), 𝐷(𝑄| 𝑃 ≤Δ . July 28, 2015 CMU: Communication Amid Uncertainty

Dictionary = Shared Randomness? Modelling the dictionary: What should it be? Simplifying assumption – it is shared randomness, so … Assume sender and receiver have some shared randomness 𝑅 and 𝑋, 𝑃, 𝑄 independent of 𝑅. 𝑌 = 𝐸(𝑃,𝑋,𝑅) 𝑋 = 𝐷(𝑄,𝑌,𝑅) Want ∀𝑋, Pr 𝑅 𝑋 = 𝑋 ≥1 −𝜖 July 28, 2015 CMU: Communication Amid Uncertainty

Solution (variant of Arith. Coding) Use 𝑅 to define sequences 𝑅 1 1 , 𝑅 1 2 , 𝑅 1 3 , … 𝑅 2 1 , 𝑅 2 2 , 𝑅 2 3 , … … 𝑅 𝑁 1 , 𝑅 𝑁 2 , 𝑅 𝑁 3 , …. 𝐸 Δ 𝑃,𝑥,𝑅 = 𝑅 𝑥 1…𝐿 , where 𝐿 chosen s.t. ∀𝑧≠𝑥 Either 𝑅 𝑧 1…𝐿 ≠ 𝑅 𝑥 1…𝐿 Or 𝑃 𝑧 < 𝑃 𝑥 4 Δ 𝐷 Δ 𝑄,𝑦,𝑅 = argmax 𝑥 {𝑄 𝑥 } among 𝑥 ∈ 𝑧 𝑅 𝑧 [1…𝐿]=𝑦 July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Performance Obviously decoding always correct. Easy exercise: Exp 𝑋 𝐸 𝑃,𝑋 =𝐻 𝑃 +2 Δ Limits: No scheme can achieve 1−𝜖 ⋅[𝐻 𝑃 +Δ] Can reduce randomness needed. July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Implications Reflects the tension between ambiguity resolution and compression. Larger the ((estimated) gap in context), larger the encoding length. Entropy is still a valid measure! Coding scheme reflects the nature of human communication (extend messages till they feel unambiguous). The “shared randomness’’ assumption A convenient starting point for discussion But is dictionary independent of context? This is problematic. July 28, 2015 CMU: Communication Amid Uncertainty

Deterministic Compression: Challenge Say Alice and Bob have rankings of N players. Rankings = bijections 𝜋, 𝜎 : 𝑁 → 𝑁 𝜋 𝑖 = rank of i th player in Alice’s ranking. Further suppose they know rankings are close. ∀ 𝑖∈ 𝑁 : 𝜋 𝑖 −𝜎 𝑖 ≤2. Bob wants to know: Is 𝜋 −1 1 = 𝜎 −1 1 How many bits does Alice need to send (non-interactively). With shared randomness – 𝑂(1) Deterministically? 𝑂 1 ? 𝑂( log 𝑁 )?𝑂( log log log 𝑁)? With Elad Haramaty: 𝑂( log ∗ 𝑛 ) July 28, 2015 CMU: Communication Amid Uncertainty

Part 2: Imperfectly Shared Randomness July 28, 2015 CMU: Communication Amid Uncertainty

Model: Imperfectly Shared Randomness Alice ←𝑟 ; and Bob ←𝑠 where 𝑟,𝑠 = i.i.d. sequence of correlated pairs 𝑟 𝑖 , 𝑠 𝑖 𝑖 ; 𝑟 𝑖 , 𝑠 𝑖 ∈{−1,+1};𝔼 𝑟 𝑖 =𝔼 𝑠 𝑖 =0;𝔼 𝑟 𝑖 𝑠 𝑖 =𝜌≥0 . Notation: 𝑖𝑠 𝑟 𝜌 (𝑓) = cc of 𝑓 with 𝜌-correlated bits. 𝑐𝑐(𝑓): Perfectly Shared Randomness cc. 𝑝𝑟𝑖𝑣 𝑓 : cc with PRIVate randomness Starting point: for Boolean functions 𝑓 𝑐𝑐 𝑓 ≤𝑖𝑠 𝑟 𝜌 𝑓 ≤𝑝𝑟𝑖𝑣 𝑓 ≤𝑐𝑐 𝑓 + log 𝑛 What if 𝑐𝑐 𝑓 ≪log 𝑛? E.g. 𝑐𝑐 𝑓 =𝑂(1) =𝑖𝑠 𝑟 1 𝑓 =𝑖𝑠 𝑟 0 𝑓 𝜌≤𝜏⇒ 𝑖𝑠 𝑟 𝜌 𝑓 ≥𝑖𝑠 𝑟 𝜏 𝑓 July 28, 2015 CMU: Communication Amid Uncertainty

Imperfectly Shared Randomness: Results Model first studied by [Bavarian,Gavinsky,Ito’14] (“Independently and earlier”). Their focus: Simultaneous Communication; general models of correlation. They show 𝑖𝑠𝑟 Equality =𝑂 1 (among other things) Our Results: Generally: 𝑐𝑐 𝑓 ≤𝑘⇒𝑖𝑠𝑟 𝑓 ≤ 2 𝑘 Converse: ∃𝑓 with 𝑐𝑐 𝑓 ≤𝑘 & 𝑖𝑠𝑟 𝑓 ≥ 2 𝑘 July 28, 2015 CMU: Communication Amid Uncertainty

Aside: Easy CC Problems Equality testing: 𝐸𝑄 𝑥,𝑦 =1⇔𝑥=𝑦; 𝐶𝐶 𝐸𝑄 =𝑂 1 Hamming distance: 𝐻 𝑘 𝑥,𝑦 =1⇔Δ 𝑥,𝑦 ≤𝑘; 𝐶𝐶 𝐻 𝑘 =𝑂(𝑘 log 𝑘 ) [Huang etal.] Small set intersection: ∩ 𝑘 𝑥,𝑦 =1⇔ wt 𝑥 , wt 𝑦 ≤𝑘 & ∃𝑖 𝑠.𝑡. 𝑥 𝑖 = 𝑦 𝑖 =1. 𝐶𝐶 ∩ 𝑘 =𝑂(𝑘) [Håstad Wigderson] Gap (Real) Inner Product: 𝑥,𝑦∈ ℝ 𝑛 ; 𝑥 2 , 𝑦 2 =1; 𝐺𝐼 𝑃 𝑐,𝑠 𝑥,𝑦 =1 if 𝑥,𝑦 ≥𝑐; =0 if 𝑥,𝑦 ≤𝑠; 𝐶𝐶 𝐺𝐼 𝑃 𝑐,𝑠 =𝑂 1 𝑐−𝑠 2 ; [Alon, Matias, Szegedy] Protocol: Fix ECC 𝐸: 0,1 𝑛 → 0,1 𝑁 ; Shared randomness: 𝑖← 𝑁 ; Exchange 𝐸 𝑥 𝑖 , 𝐸 𝑦 𝑖 ; Accept iff 𝐸 𝑥 𝑖 =𝐸 𝑦 𝑖 . 𝑝𝑜𝑙𝑦 𝑘 Protocol: Use common randomness to hash 𝑛 →[ 𝑘 2 ] 𝑥=( 𝑥 1 , …, 𝑥 𝑛 ) 𝑦= 𝑦 1 , …, 𝑦 𝑛 〈𝑥,𝑦〉≜ 𝑖 𝑥 𝑖 𝑦 𝑖 Main Insight: If 𝐺←𝑁 0,1 𝑛 , then 𝔼 𝐺,𝑥 ⋅ 𝐺,𝑦 =〈𝑥,𝑦〉 [Ghazi,Kamath,S’15]: Roughly essence of perm. inv. functions July 28, 2015 CMU: Communication Amid Uncertainty

Equality Testing (our proof) Key idea: Think inner products. Encode 𝑥↦𝑋=𝐸(𝑥);𝑦↦𝑌=𝐸 𝑦 ;𝑋,𝑌∈ −1,+1 𝑁 𝑥=𝑦⇒ 〈𝑋,𝑌〉=𝑁 𝑥≠𝑦⇒ 〈𝑋,𝑌〉≤𝑁/2 Estimating inner products: Building on sketching protocols … Alice: Picks Gaussians 𝐺 1 ,… 𝐺 𝑡 ∈ ℝ 𝑁 , Sends 𝑖∈ 𝑡 maximizing 𝐺 𝑖 ,𝑋 to Bob. Bob: Accepts iff 𝐺′ 𝑖 ,𝑌 ≥0 Analysis: 𝑂 𝜌 (1) bits suffice if 𝐺 ≈ 𝜌 𝐺′ Gaussian Protocol July 28, 2015 CMU: Communication Amid Uncertainty

General One-Way Communication Idea: All communication ≤ Inner Products (For now: Assume one-way-cc 𝑓 ≤ 𝑘) For each random string 𝑅 Alice’s message = 𝑖 𝑅 ∈ 2 𝑘 Bob’s output = 𝑓 𝑅 ( 𝑖 𝑅 ) where 𝑓 𝑅 : 2 𝑘 → 0,1 W.p. ≥ 2 3 over 𝑅, 𝑓 𝑅 𝑖 𝑅 is the right answer. July 28, 2015 CMU: Communication Amid Uncertainty

General One-Way Communication For each random string 𝑅 Alice’s message = 𝑖 𝑅 ∈ 2 𝑘 Bob’s output = 𝑓 𝑅 ( 𝑖 𝑅 ) where 𝑓 𝑅 : 2 𝑘 → 0,1 W.p. ≥ 2 3 , 𝑓 𝑅 𝑖 𝑅 is the right answer. Vector representation: 𝑖 𝑅 ↦ 𝑥 𝑅 ∈ 0,1 2 𝑘 (unit coordinate vector) 𝑓 𝑅 ↦ 𝑦 𝑅 ∈ 0,1 2 𝑘 (truth table of 𝑓 𝑅 ). 𝑓 𝑅 𝑖 𝑅 =〈 𝑥 𝑅 , 𝑦 𝑅 〉; Acc. Prob. ∝ 𝑋,𝑌 ;𝑋= 𝑥 𝑅 𝑅 ;𝑌= 𝑦 𝑅 𝑅 Gaussian protocol estimates inner products of unit vectors to within ±𝜖 with 𝑂 𝜌 1 𝜖 2 communication. July 28, 2015 CMU: Communication Amid Uncertainty

Two-way communication Still decided by inner products. Simple lemma: ∃ 𝐾 𝐴 𝑘 , 𝐾 𝐵 𝑘 ⊆ ℝ 2 𝑘 convex, that describe private coin k-bit comm. strategies for Alice, Bob s.t. accept prob. of 𝜋 𝐴 ∈ 𝐾 𝐴 𝑘 , 𝜋 𝐵 ∈ 𝐾 𝐵 𝑘 equals 〈 𝜋 𝐴 , 𝜋 𝐵 〉 Putting things together: Theorem: 𝑐𝑐 𝑓 ≤𝑘⇒𝑖𝑠𝑟 𝑓 ≤ 𝑂 𝜌 ( 2 𝑘 ) July 28, 2015 CMU: Communication Amid Uncertainty

Part 3: Uncertain Functionality July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Model Alice knows 𝑔≈𝑓; Bob wishes to compute 𝑓(𝑥,𝑦) Alice, Bob given 𝑔,𝑓 explicitly. (Input size ~ 2 𝑛 ) Questions: What is ≈? Is it reasonable to expect to compute 𝑓 𝑥,𝑦 ? E.g., 𝑓 𝑥,𝑦 = 𝑓 ′ 𝑥 ? Can’t compute 𝑓 𝑥,𝑦 without communicating 𝑥 Answers: Assume 𝑥,𝑦∼ 0,1 𝑛 × 0,1 𝑛 uniformly. 𝑓 ≈ 𝛿 𝑔 if 𝛿 𝑓,𝑔 ≤𝛿. Suffices to compute ℎ 𝑥,𝑦 for ℎ ≈ 𝜖 𝑓 July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Results - 1 Thm [Ghazi,Komargodski,Kothari,S.]: ∀𝜖>0, ∃𝛿>0 s.t. If 𝑓 has one-way communication 𝑘, then in the 𝜖,𝛿 −uncertain model it has communication complexity 𝑂 𝑘 . Main Idea: Canonical protocol for 𝑓: Alice + Bob share random 𝑥 1 , … 𝑥 𝑚 ∈ 0,1 𝑛 . Alice sends 𝑓 𝑥 1 , …, 𝑓 𝑥 𝑚 to Bob. Protocol used previously … but not as “canonical”. Canonical protocol robust when 𝑓≈𝑔. July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Results - 2 Can extend model to 𝑥,𝑦 ∼𝜇 for arbitrary distribution 𝜇 Results: If 𝜇 is product distribution (𝑥⊥𝑦) then results extend. Else Upper bound: Multiplicative factor of 𝐼(𝑥,𝑦) Lower bound: Some blowup necessary ∃𝜇 and dist. on pairs of functions (𝑓,𝑔) of constant comm. complexity; but computing 𝑔(𝑥,𝑦) in the uncertain model costs Ω 𝑛 bits. Open: Significance of above? July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Conclusions Context Important: New layer of uncertainty. New notion of scale (context LARGE) Many open directions+questions July 28, 2015 CMU: Communication Amid Uncertainty

CMU: Communication Amid Uncertainty Thank You! July 28, 2015 CMU: Communication Amid Uncertainty