The Price of Uncertainty in Communication Brendan Juba (Washington U., St. Louis) with Mark Braverman (Princeton)

Slides:



Advertisements
Similar presentations
1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Advertisements

Limitations of Quantum Advice and One-Way Communication Scott Aaronson UC Berkeley IAS Useful?
Xiaoming Sun Tsinghua University David Woodruff MIT
Ulams Game and Universal Communications Using Feedback Ofer Shayevitz June 2006.
Detection of Algebraic Manipulation with Applications to Robust Secret Sharing and Fuzzy Extractors Ronald Cramer, Yevgeniy Dodis, Serge Fehr, Carles Padro,
Approximate List- Decoding and Hardness Amplification Valentine Kabanets (SFU) joint work with Russell Impagliazzo and Ragesh Jaiswal (UCSD)
Of 24 11/20/2012TIFR: Deterministic Communication Amid Uncertainty1 ( Deterministic ) Communication amid Uncertainty Madhu Sudan Microsoft, New England.
The Communication Complexity of Approximate Set Packing and Covering
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
15-583:Algorithms in the Real World
Foundations of Cryptography Lecture 10 Lecturer: Moni Naor.
QuickSort Average Case Analysis An Incompressibility Approach Brendan Lucier August 2, 2005.
Greedy Algorithms Amihood Amir Bar-Ilan University.
Chapter 10 Shannon’s Theorem. Shannon’s Theorems First theorem:H(S) ≤ L n (S n )/n < H(S) + 1/n where L n is the length of a certain code. Second theorem:
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Foundations of Cryptography Lecture 4 Lecturer: Moni Naor.
Data Compression.
Entropy and Shannon’s First Theorem
Chapter 6 Information Theory
ENGS Lecture 8 ENGS 4 - Lecture 8 Technology of Cyberspace Winter 2004 Thayer School of Engineering Dartmouth College Instructor: George Cybenko,
CPSC 411, Fall 2008: Set 4 1 CPSC 411 Design and Analysis of Algorithms Set 4: Greedy Algorithms Prof. Jennifer Welch Fall 2008.
6/20/2015List Decoding Of RS Codes 1 Barak Pinhas ECC Seminar Tel-Aviv University.
Lossless data compression Lecture 1. Data Compression Lossless data compression: Store/Transmit big files using few bytes so that the original files.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
Data Structures – LECTURE 10 Huffman coding
CS151 Complexity Theory Lecture 10 April 29, 2004.
Variable-Length Codes: Huffman Codes
CSI Uncertainty in A.I. Lecture 201 Basic Information Theory Review Measuring the uncertainty of an event Measuring the uncertainty in a probability.
CPSC 411, Fall 2008: Set 4 1 CPSC 411 Design and Analysis of Algorithms Set 4: Greedy Algorithms Prof. Jennifer Welch Fall 2008.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 13 June 22, 2005
Noise, Information Theory, and Entropy
Noise, Information Theory, and Entropy
Huffman Codes Message consisting of five characters: a, b, c, d,e
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Noiseless Coding. Introduction Noiseless Coding Compression without distortion Basic Concept Symbols with lower probabilities are represented by the binary.
A Few Simple Applications to Cryptography Louis Salvail BRICS, Aarhus University.
1 Information and interactive computation January 16, 2012 Mark Braverman Computer Science, Princeton University.
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Compression.  Compression ratio: how much is the size reduced?  Symmetric/asymmetric: time difference to compress, decompress?  Lossless; lossy: any.
Linawati Electrical Engineering Department Udayana University
Lossless Compression CIS 465 Multimedia. Compression Compression: the process of coding that will effectively reduce the total number of bits needed to.
Prof. Amr Goneid, AUC1 Analysis & Design of Algorithms (CSCE 321) Prof. Amr Goneid Department of Computer Science, AUC Part 8. Greedy Algorithms.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Coding Theory Efficient and Reliable Transfer of Information
1 Private codes or Succinct random codes that are (almost) perfect Michael Langberg California Institute of Technology.
Of 22 10/30/2015WUSTL: Uncertain Communication1 Communication Amid Uncertainty Madhu Sudan Harvard Based on Juba, S. (STOC 2008, ITCS 2011) Juba, S. (STOC.
Basic Concepts of Encoding Codes and Error Correction 1.
Data Stream Algorithms Lower Bounds Graham Cormode
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Compression for Fixed-Width Memories Ori Rottenstriech, Amit Berman, Yuval Cassuto and Isaac Keslassy Technion, Israel.
SEAC-3 J.Teuhola Information-Theoretic Foundations Founder: Claude Shannon, 1940’s Gives bounds for:  Ultimate data compression  Ultimate transmission.
Ch4. Zero-Error Data Compression Yuan Luo. Content  Ch4. Zero-Error Data Compression  4.1 The Entropy Bound  4.2 Prefix Codes  Definition and.
Communication Amid Uncertainty
Huffman Codes ASCII is a fixed length 7 bit code that uses the same number of bits to define each character regardless of how frequently it occurs. Huffman.
Information Complexity Lower Bounds
Sublinear-Time Error-Correction and Error-Detection
The Greedy Method and Text Compression
Communication Amid Uncertainty
The Greedy Method and Text Compression
Communication Amid Uncertainty
Communication Amid Uncertainty
Lecture 9 Greedy Strategy
Uncertain Compression
Distributed Compression For Binary Symetric Channels
Imperfectly Shared Randomness
Hash Functions Motivation Hash Functions: collision, pre-images SHA-1
Information-Theoretic Security
Communication Amid Uncertainty
One Way Functions Motivation Complexity Theory Review, Motivation
Presentation transcript:

The Price of Uncertainty in Communication Brendan Juba (Washington U., St. Louis) with Mark Braverman (Princeton)

≈ ≈ SINCE WE ALL AGREE ON A PROB. DISTRIBUTION OVER WHAT I MIGHT SAY, I CAN COMPRESS IT TO: “THE 9,232,142,124,214,214,123,845 TH MOST LIKELY MESSAGE. THANK YOU!”

1.Encodings and communication across different priors 2.Near-optimal lower bounds for different priors coding 3

Coding schemes 4 BirdChicken Cat DinnerPetLambDuckCowDog “MESSAGES” “ENCODINGS”

Communication model 5 RECALL: (, CAT)  E

Ambiguity 6 BirdChicken Cat DinnerPetLambDuckCowDog

Prior distributions 7 BirdChicken Cat DinnerPetLambDuckCowDog Decode to a maximum likelihood message

Source coding (compression) Assume encodings are binary strings Given a prior distribution P, message m, choose minimum length encoding that decodes to m. 8 FOR EXAMPLE, HUFFMAN CODES AND SHANNON- FANO (ARITHMETIC) CODES NOTE: THE ABOVE SCHEMES DEPEND ON THE PRIOR.

9 SUPPOSE ALICE AND BOB SHARE THE SAME ENCODING SCHEME, BUT DON’T SHARE THE SAME PRIOR… P P Q Q CAN THEY COMMUNICATE?? HOW EFFICIENTLY??

10 THE CAT. THE ORANGE CAT. THE ORANGE CAT WITHOUT A HAT.

Closeness and communication Priors P and Q are α-close (α ≥ 1) if for every message m, αP(m) ≥ Q(m) and αQ(m) ≥ P(m) Disambiguation and closeness together suffice for communication: If for every m’≠m, P[m|e] > α 2 P[m’|e], then: Q[m|e] ≥ 1 / α P[m|e] > αP[m’|e] ≥ Q[m’|e] 11 SO, IF ALICE SENDS e THEN MAXIMUM LIKELIHOOD DECODING GIVES BOB m AND NOT m’… “α 2 -disambiguated”

Construction of a coding scheme (J-Kalai-Khanna-Sudan’11, Inspired by B-Rao’11) Pick an infinite random string R m for each m, Put (m,e)  E ⇔ e is a prefix of R m. Alice encodes m by sending prefix of R m s.t. m is α 2 -disambiguated under P. 12 Gives an expected encoding length of at most H(P) + 2log α + 2

Remark Mimicking the disambiguation property of natural language provided an efficient strategy for communication. 13

1.Encodings and communication across different priors 2.Near-optimal lower bounds for different priors coding 14

Our results 1.The JKKS’11/BR’11 encoding is near optimal – H(P) + 2log α – 3log log α – O(1) bits necessary (cf. achieved H(P) + 2log α + 2 bits) 2.Analysis of positive-error setting [Haramaty-Sudan’14] : If incorrect decoding w.p. ε is allowed— – Can achieve H(P) + log α + log 1/ε bits – H(P) + log α + log 1/ε – (9/2)log log α – O(1) bits necessary for ε > 1/α 15

An ε-error coding scheme. (Inspired by J-Kalai-Khanna-Sudan’11, B-Rao’11) Pick an infinite random string R m for each m, Put (m,e)  E ⇔ e is a prefix of R m. Alice encodes m by sending the prefix of R m of length log 1/P(m) + log α + log 1/ε 16

Analysis Claim. m is decoded correctly w.p. 1-ε Proof. There are at most 1/Q(m) messages with Q-probability greater than Q(m) ≥ P(m)/α. The probability that R m’ for any one of these m’ agrees with the first log 1/P(m) + log α + log 1/ε ≥ log 1/Q(m)+log 1/ε bits of R m is at most εQ(m). By a union bound, the probability that any of these agree with R m (and hence could be wrongly chosen) is at most ε. 17

Length lower bound 1—reduction to deterministic encodings Min-max Theorem: it suffices to exhibit a distribution over priors for which deterministic encodings must be long 18

Length lower bound 2—hard priors 19 log. prob. ≈0 -log α -2log α m*m* S Lemma 1: H(P) = O(1) α- close Lemma 2 α2α2 α

Length lower bound 3— short encodings have collisions Encodings of expected length < 2log α – 3log log α encode m 1 ≠ m 2 identically with nonzero prob. With nonzero probability over choice of P & Q, m 1,m 2 ∈ S and m* ∈ {m 1,m 2 } Decoding error with nonzero probability ☞ Errorless encodings have expected length ≥ 2log α-3log log α = H(P)+2log α-3log log α-O(1) 20

Length lower bound 4— very short encodings often collide If the encoding has expected length < log α + log 1/ε – (9/2)log log α m* collides with ∼ (ε log α)∙α other messages Probability that our α draws for S miss all of these messages is < 1-2ε Decoding error with probability > ε ☞ Error-ε encodings have expected length ≥ H(P) + log α + log 1/ε – (9/2)log log α – O(1) 21

22 Recap. We saw a variant of source coding for which (near-)optimal solutions resemble natural languages in interesting ways.

23 The problem. Design a coding scheme E so that for any sender and receiver with α-close prior distributions, the communication length is minimized. (In expectation w.r.t. sender’s distribution) Questions?