Communication Complexity, Information Complexity and Applications to Privacy Toniann Pitassi University of Toronto.

Slides:



Advertisements
Similar presentations
1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Advertisements

Optimal Space Lower Bounds for All Frequency Moments David Woodruff MIT
Quantum Lower Bounds The Polynomial and Adversary Methods Scott Aaronson September 14, 2001 Prelim Exam Talk.
Numerical Linear Algebra in the Streaming Model Ken Clarkson - IBM David Woodruff - IBM.
Optimal Space Lower Bounds for all Frequency Moments David Woodruff Based on SODA 04 paper.
The Average Case Complexity of Counting Distinct Elements David Woodruff IBM Almaden.
Optimal Bounds for Johnson- Lindenstrauss Transforms and Streaming Problems with Sub- Constant Error T.S. Jayram David Woodruff IBM Almaden.
Xiaoming Sun Tsinghua University David Woodruff MIT
Tight Lower Bounds for the Distinct Elements Problem David Woodruff MIT Joint work with Piotr Indyk.
Truthful Mechanisms for Combinatorial Auctions with Subadditive Bidders Speaker: Shahar Dobzinski Based on joint works with Noam Nisan & Michael Schapira.
Separating Deterministic from Randomized Multiparty Communication Complexity Joint work with Paul Beame (University of Washington) Matei David (University.
The Communication Complexity of Approximate Set Packing and Covering
An Approximate Truthful Mechanism for Combinatorial Auctions An Internet Mathematics paper by Aaron Archer, Christos Papadimitriou, Kunal Talwar and Éva.
Gillat Kol (IAS) joint work with Ran Raz (Weizmann + IAS) Interactive Channel Capacity.
Gillat Kol joint work with Ran Raz Competing Provers Protocols for Circuit Evaluation.
Seminar In Game Theory Algorithms, TAU, Agenda  Introduction  Computational Complexity  Incentive Compatible Mechanism  LP Relaxation & Walrasian.
Seminar in Foundations of Privacy 1.Adding Consistency to Differential Privacy 2.Attacks on Anonymized Social Networks Inbal Talgam March 2008.
On the tightness of Buhrman- Cleve-Wigderson simulation Shengyu Zhang The Chinese University of Hong Kong On the relation between decision tree complexity.
CS151 Complexity Theory Lecture 7 April 20, 2004.
Limitations of VCG-Based Mechanisms Shahar Dobzinski Joint work with Noam Nisan.
Approximate Privacy: Foundations and Quantification Michael Schapira (Yale and UC Berkeley) Joint work with Joan Feigenbaum (Yale) and Aaron D. Jaggard.
Sketching in Adversarial Environments Or Sublinearity and Cryptography 1 Moni Naor Joint work with: Ilya Mironov and Gil Segev.
On Gallager’s problem: New Bounds for Noisy Communication. Navin Goyal & Mike Saks Joint work with Guy Kindler Microsoft Research.
An Introduction to Black-Box Complexity
1 Approximate Privacy: Foundations and Quantification Joan Feigenbaum DIMACS; November 20, 2009 Joint work with A. D. Jaggard.
1 Streaming Computation of Combinatorial Objects Ziv Bar-Yossef U.C. Berkeley Omer Reingold AT&T Labs – Research Ronen.
1 Approximate Privacy: Foundations and Quantification Joan Feigenbaum Northwest Univ.; May 20, 2009 Joint work with A.
. PGM 2002/3 – Tirgul6 Approximate Inference: Sampling.
Shengyu Zhang The Chinese University of Hong Kong.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 13 June 22, 2005
Lower Bounds for Massive Dataset Algorithms T.S. Jayram (IBM Almaden) IIT Kanpur Workshop on Data Streams.
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Multiplicative Weights Algorithms CompSci Instructor: Ashwin Machanavajjhala 1Lecture 13 : Fall 12.
Complexity Theory Lecture 2 Lecturer: Moni Naor. Recap of last week Computational Complexity Theory: What, Why and How Overview: Turing Machines, Church-Turing.
Secure Computation of the k’th Ranked Element Gagan Aggarwal Stanford University Joint work with Nina Mishra and Benny Pinkas, HP Labs.
Ragesh Jaiswal Indian Institute of Technology Delhi Threshold Direct Product Theorems: a survey.
1 Information and interactive computation January 16, 2012 Mark Braverman Computer Science, Princeton University.
Quantum Computing MAS 725 Hartmut Klauck NTU TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A.
Streaming Algorithms Piotr Indyk MIT. Data Streams A data stream is a sequence of data that is too large to be stored in available memory Examples: –Network.
Preference elicitation Communicational Burden by Nisan, Segal, Lahaie and Parkes October 27th, 2004 Jella Pfeiffer.
Communication Complexity Rahul Jain Centre for Quantum Technologies and Department of Computer Science National University of Singapore. TexPoint fonts.
Private Approximation of Search Problems Amos Beimel Paz Carmi Kobbi Nissim Enav Weinreb (Technion)
Information Theory for Data Streams David P. Woodruff IBM Almaden.
Personalized Social Recommendations – Accurate or Private? A. Machanavajjhala (Yahoo!), with A. Korolova (Stanford), A. Das Sarma (Google) 1.
Communication vs. Computation S Venkatesh Univ. Victoria Presentation by Piotr Indyk (MIT) Kobbi Nissim Microsoft SVC Prahladh Harsha MIT Joe Kilian NEC.
Foundations of Privacy Lecture 5 Lecturer: Moni Naor.
Asymmetric Communication Complexity And its implications on Cell Probe Complexity Slides by Elad Verbin Based on a paper of Peter Bro Miltersen, Noam Nisan,
Fall 2013 CMU CS Computational Complexity Lectures 8-9 Randomness, communication, complexity of unique solutions These slides are mostly a resequencing.
Umans Complexity Theory Lectures Lecture 7b: Randomization in Communication Complexity.
Massive Data Sets and Information Theory Ziv Bar-Yossef Department of Electrical Engineering Technion.
Data Stream Algorithms Lower Bounds Graham Cormode
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Tight Bound for the Gap Hamming Distance Problem Oded Regev Tel Aviv University TexPoint fonts used in EMF. Read the TexPoint manual before you delete.
Gillat Kol (IAS) joint work with Anat Ganor (Weizmann) Ran Raz (Weizmann + IAS) Exponential Separation of Information and Communication.
Approximation Algorithms for Combinatorial Auctions with Complement-Free Bidders Speaker: Shahar Dobzinski Joint work with Noam Nisan & Michael Schapira.
1 Differential Privacy Cynthia Dwork Mamadou H. Diallo.
Hartmut Klauck Centre for Quantum Technologies Nanyang Technological University Singapore.
The Message Passing Communication Model David Woodruff IBM Almaden.
Information Complexity Lower Bounds
New Characterizations in Turnstile Streams with Applications
Hans Bodlaender, Marek Cygan and Stefan Kratsch
Understanding Generalization in Adaptive Data Analysis
On Communication Protocols that Compute Almost Privately
Differential Privacy in Practice
CS 154, Lecture 6: Communication Complexity
Linear sketching with parities
Linear sketching with parities
Imperfectly Shared Randomness
CS21 Decidability and Tractability
Shengyu Zhang The Chinese University of Hong Kong
Presentation transcript:

Communication Complexity, Information Complexity and Applications to Privacy Toniann Pitassi University of Toronto

2-Party Communication Complexity [Yao] 2-party communication: each party has a dataset. Goal is to compute a function f(D A,D B ) m1m1 m2m2 m3m3 m k-1 mkmk DADA x1x1 x2x2  xnxn DBDB y1y1 y2y2  ymym f(D A,D B ) Communication complexity of a protocol for f is the number of bits exchanged between A and B. In this talk, all protocols are assumed to be randomized.

Deterministic Protocols A deterministic protocol Π specifies: – Function of board contents: if the protocol is over if YES, the output if NO, which player writes next – Function of board contents and input available to player P: what P writes Cost of Π = max number of bits written on the board over all inputs

Randomized Protocols In a randomized protocol Π, what player P writes is also a function of the (private and/or public) random string available to P Protocol allowed to err with probability ε over choice of random strings The cost of Π = max number of bits written on the board, over inputs and random strings

Communication Complexity Focus on randomized communication complexity: CC(F,ε) = the communication cost of computing F with error ε. A distributional flavor of randomized communication complexity: CC(F,μ,ε) = the communication cost of computing F with error ε with respect to μ. Yao’s minimax: CC(F,ε)=max μ CC(F,μ,ε). 5

Stunning variety of applications of CC Lower Bounds 1.Lower Bounds for Streaming Algorithms 2.Data Structure Lower Bounds 3.Proof Complexity Lower Bounds 4.Game Theory 5.Circuit Complexity Lower Bounds 6.Quantum Computation 7.Differential Privacy 8.……

2-Party Information Complexity 2-party communication: each party has a dataset. Goal is to compute a function f(D A,D B ) m1m1 m2m2 m3m3 m k-1 mkmk DADA x1x1 x2x2  xnxn DBDB y1y1 y2y2  ymym f(D A,D B ) Information complexity of a protocol for f is the amount of information the players reveal to each other / or to an eavesdropper (Eve)

Information Complexity [Chakrabarti,Shi,Wirth,Yao ‘01], [Bar-Yossef,Jayram,Kumar,Sivakumar ‘04] Entropy: H(X) = Σ x p(x) log (1/p(x) Conditional entropy: H(X|Y) = Σ y H(X|Y=y) p(Y=y) Mutual Information: I(X;Y) = H(X) - H(X|Y) External IC: information about XY revealed to Eve IC ext (π,μ) = I(XY;π) IC ext (f,μ,ε) = max π IC ext (π,μ) Internal IC: information revealed to Alice and Bob IC int (π,μ) = I(X;π|Y) + I(Y;π|X) IC int (f,μ,ε) = max π IC int (π,μ)

Why study information complexity? Intrinsically interesting quantity Related to longstanding questions in complexity theory (direct sum conjecture) Very useful when studying privacy, and quantum computation

Simple Facts about Information Complexity External information cost is greater than internal: IC ext (π,μ) ≥ IC int (π,μ) IC ext (π) = I(XY;π) = I(X;π) + I(Y;π | X) ≥ I(X;π|Y) + I(Y; π | X) = IC int (π) Information complexity lower bounds imply Communication Complexity lower bounds: CC(f,μ,ε) ≥ IC ext (f,μ,ε) ≥ IC int (f,μ,ε)

Do CC Lower Bounds imply IC Lower Bounds? (i.e., CC=IC?) For constant-round protocols, IC and CC are basically equal [CSWY, JRS] Open for general protocols. Significant step for general case by [BBCR]

Compressing Interactive Communication [Barak,Braverman,Chen,Rao] Theorem 1 For any distribution μ, any C-bit protocol of internal IC I can be simulated by a new protocol using O( √( CI) logC) bits. Theorem 2 For any product distribution μ, any C-bit protocol of internal IC I can be simulated by a protocol using O(I logC) bits.

Connection to the Direct Sum Problem Does it take m times the amount of resources to solve m instances? Direct Sum Question for CC: CC(f m ) ≥ m CC(f) for every f and every distribution? - Each copy should have error ε For search problems, the direct sum problem is equivalent to separating NC 1 from P !

Connection to the Direct Sum Problem, 2 The direct sum property holds for information complexity: Lemma [Direct Sum for IC]: IC(f m ) ≥ m IC(f) Best general direct sum theorem known for cc: Theorem [Barak,Braverman,Chen,Rao]: CC(f m ) ≥ √ m CC(f) ignoring polylog factors The direct sum property for cc is equivalent to IC=CC! Theorem [Braverman,Rao]: IC(f, μ,ε ) = lim n  ∞ CC(F n, μ n, ε )/n

Methods for Proving CC and IC Lower Bounds Jain and Klauck initiated the formal study of CC lower bound methods: all formalizable as solutions to (different) LPs Discrepancy Method, Smooth Discrepancy Method Rectangle Bound, Smooth Rectangle Bound Partition Bound

The Partition Bound [Jain, Klauck] Min Σ z,R w z,R ∀ (x,y) Σ R, (x,y) ϵ R w f(x,y),R ≥ 1-ε ∀ (x,y) Σ R, (x,y) in R Σ z w z,R = 1 ∀ z,R w z,R ≥ 0

Relationships The Partition bound is greater than or equal to all known CC lower bounds methods, including: Discrepancy Generalized Discrepancy Rectangle Smooth Rectangle

[KLLR] define the relaxed partition bound. The relaxed partition bound is greater than or equal to all known CC lower bound methods (except the partition bound). They show that the relaxed Partition bound is equivalent to designing a zero-communication protocol with error exp(-I) Given a protocol for f with IC int = I, they construct a zero-communication protocol st (i) non-abort probability is exp(-I), and (ii) if it does not abort, it computes f correctly whp All known CC Lower Bound Methods Imply IC Lower Bounds! [Kerenidis, Laplante, Lerays, Roland Xiao ‘12]

Applications of Information Complexity Differential Privacy PAR

Applications of Information Complexity Differential Privacy PAR

Differential Privacy: The Basic Scenario [Dwork, McSherry, Nissim, Smith 06] Database with rows x 1..x n Each row corresponds to an individual in the database Columns correspond to fields, such as “name”, “zip code”; some fields contain sensitive information. Goal: Compute and release information about a sensitive database without revealing information about any individual Sanitizer Output Data

Differential Privacy [Dwork,McSherry,Nissim,Smith 2006] Y Pr [response] ratio bounded Q = space of queries; Y = output space; X = row space Mechanism M: X n x Q  Y is  -differentially private if: for all q in Q, for all adjacent x, x’ in X n, the distributions M(x,q), M(x’,q) are similar: ∀ y in Y, q in Q: e - ≤ Pr[M(x,q) =y] ≤ e ε Pr[M(x’,q)=y] Note: Randomness is crucial

23 Achieving DP: Add Laplacian Noise  f = max D,D’ |f(D) – f(D’)| 0 b2b3b4b5b-b-2b-3b-4b Theorem: To achieve  -differential privacy, add symmetric noise [Lap(b)] with b =  f/ . P(y) ∽ exp(-|y - q(x)|/b) = exp( - | y – q(x’)|  /  f ) Pr [M(x, q) = y] Pr [(M(x’, q) = y] exp( - | y – q(x)|  /  f ) ≤ exp(  ).

Differentially Private Communication Complexity: A Distributed View Andrews,Mironov,P,Reingold,Talwar,Vadhan Goal: compute a joint function while maintaining privacy for any individual, with respect to both the outside world and the other database owners. Multiple databases, each with private data. D1 D2 D3 D4D5 F(D1,D2,..,D5)

2-Party Differentially Private CC 2-party (& multiparty) DP privacy: each party has a dataset; want to compute a joint function f(D A,D B ) m1m1 m2m2 m3m3 m k-1 mkmk DADA x1x1 x2x2  xnxn DBDB y1y1 y2y2  ymym Z A  f(D A,D B ) Z B  f(D A,D B ) A’s view should be a differentially private function of D B (even if A deviates from protocol), and vice-versa

Two-Party Differential Privacy Let P(x,y) be a 2-party protocol. P is ε-DP if: (1) for all y, for every pair x, x’ that are neighbors, and for every transcript π, Pr[P(x,y) = π ] ≤ exp(ε) Pr[P(x’,y) = π ] (2) symmetrically, for all x, for every pair of neighbors y,y’ and for every transcript π Pr[P(x,y)=π ] ≤ exp(ε) Pr[P(x,y’) = π] Privacy and accuracy are the important parameters

Examples 1.Ones(x,y) = the number of ones in xy Ones( , ) = 8. CC(Ones) = logn. There is a low error DP protocol. 2. Hamming Distance HD(x,y) = the number of positions i where x i ≠ y i. HD( , ) = 4 CC(HD)=n. No low error DP protocol Is this a coincidence? Is there a connection between low cc and low-error DP protocols?

Information Cost and DP Protocols [McGregor, Mironov, P,Reingold,Talwar,Vadhan] Lemma. If π has ε-DP, then for every distribution μ on XY, IC(π,μ,ε) ≤ 3εn Proof sketch: For every z,z’, by ε-DP, exp(-2εn) ≤ Pr[π(z) = π]/Pr[π(z’)=π] ≤ exp(2εn) I(π(Z); X) = H(π(Z)) – H(π(Z) | Z) = Exp {z,π} log[ Pr[π(Z)=π | Z=z] / Pr[π(Z)=π] ] ≤ 2 (log ε) εn DP Partition Theorem. Let P be an ε -DP protocol for a partial function with error at most γ. Then log prt γ (f) ≤ 3 ε n

Lower Bound:Hamming Distance [McGregor, Mironov, P,Reingold,Talwar,Vadhan] Gap Hamming: GHD(x,y) = 1 if HD(x,y) > n/2 + √ n 0 if HD(x,y) < n/2 – √ n Theorem. Any ε-DP protocol for Hamming distance must incur an additive error Ω(√n). Note: This lower bound is tight. Proof sketch: [Chakrabarti-Regev 2012] prove: CC(GHD, μ,1/3) = Ω (n). Proof shows GHD has a smooth rectangle bound of 2 Ω (n). By Jain-Klauck, this implies that the partition bound for GHD is at least as large. Thus proof follows by DP Partition Theorem.

Implications of Lower bound for Hamming Distance 1. Separation between ε -DP protocols and computational ε -DP protocols [MPRV]: Hamming distance has an O(1) error computational ε -DP protocol, but any ε -DP protocol has error √ n. We also exhibit another function with linear separation. (Any ε- DP protocol has error Ω n) 2. Pan Privacy: Our lower bound for Hamming Distance implies lower bounds for pan-private streaming algorithms

Pan-Private Streaming Model [Dwork,P,Rothblum, Naor,Yekhanin] Data is a stream of items; each item belongs to a user. Sanitizer sees each item and updates internal state. Generates output at end of the stream (single pass). state Pan-Privacy: For every two adjacent streams, at any single point in time, the internal state (and final output) are differentially private.

What statistics have pan-private algorithms? We give pan-private streaming algorithms for: Stream density / number of distinct elements t-cropped mean: mean, over users, of min(t, #appearances) Fraction of users appearing exactly k times Fraction of users appearing exactly 0 times modulo k Fraction of heavy-hitters, users appearing at least k times

Pan Privacy lower bounds via ε -DP lower bounds Lower Bounds for ε-DP communication protocols imply pan privacy lower bounds for density estimation (via Hamming distance lower bound). Lower bounds also hold for multi-pass pan- private models Analogy: 2-party communication complexity lower bounds imply lower bounds in streaming model.

DP Protocols and Compression So back to Ones(x,y) and HD(x,y)...is DP the same as compressible? Theorem. [BBCR] (Low Icost implies compression) For every product distribution μ, and protocol P, there exists a protocol Q ( β- approximating P) with comm. complexity ∼ Icost μ (P) x polylog(CC(P))/ β Corollary. (DP protocols can be compressed) Let P be an ε -DP protocol P. Then there exists a protocol Q of cost 3 ε n polylog(CC(P))/ β and error β. DP almost implies low cc, except for this annoying polylog(CC(P)) factor Moreover, the low cc protocol can often be made DP (if the number of rounds is bounded.)

Differential Privacy and Compression We have seen that DP protocols have low information cost By BBCR this implies they can be compressed (and thus have low comm complexity) What about the other direction? Can functions with low cc be made DP? Yes! (with some caveats..the error is proportional not only to the cc, but also the number of rounds.) Proof uses the exponential mechanism [MT]

Applications of Information Complexity Differential Privacy PAR

37 Approximate Privacy in Mechanism Design Traditional goal of mechanism design: Incent agents to reveal private information that is needed to compute optimal results. Complementary, newly important goal: Enable agents not to reveal private information that is not needed to compute optimal results. Example (Naor-Pinkas-Sumner, EC ’99): It’s undesirable for the auctioneer to learn the winning bid in a 2 nd –price Vickrey auction.

38 Perfect Privacy [Kushilevitz ’92] Protocol P for f is perfectly private iff for all x,x’,y,y’ f(x,y)=f(x’,y’)  R(x,y)=R(x’,y’) f is perfectly privately computable iff M(f) has no forbidden submatrix f(x 1, x 2 ) = f ( x’ 1, x 2 ) = f(x’ 1, x’ 2 ) = a, but f ( x 1, x’ 2 ) ≠ a x 1 x’ 1 X 2 X’ 2

39 Example 1: Millionaires’ Problem (not perfectly privately computable) millionaire 1 millionaire 2 A(f) f(x 1, x 2 ) = 1 if x 1 ≥ x 2 ; else f(x 1, x 2 ) = 2

40 Example 2: Vickrey Auction [Brandt, Sandholm] 2, 1 winner price 2, 0 1, 0 1, 1 1, 2 2, 2 1, bidder 1 bidder R I (2, 0) The ascending-price, English auction protocol is the unique perfectly private protocol However the communication cost is exponential !!

41 Worst-case PAR [Feigenbaum, Jaggard,Schapira ‘10] Worst-case privacy approximation ratio of a protocol π for f: PAR(f, π) = max x,y | P(x,y)|/ |R(x,y)|, P(x,y): set of all pairs (x’,y’) st f(x,y)=f’(x’,y’) R(x,y): rectangle containing (x,y) induced by π Worst-case PAR of f: PAR(f) = min π PAR(f,π)

42 Average-case PAR [Feigenbaum, Jaggard, Schapira ‘10] (1) Average-case PAR of π: AvgPAR1(f,π) = log E (x,y) |P(x,y)|/|R(x,y)| AvgPAR1(f) = min π AvgPAR(f,π) (2) Alternative definition: AvgPAR2(f,π) = I(XY; π | f) = E (x,y) log |P(x,y)/|R(x,y)| AvgPAR2(f) = min π AvgPAR2(f,π) 1 is log of Expectation, 2 is Expectation of log. For boolean functions, AvgPAR2(f) is basically the same as Icost(f) (differs by at most 1).

43 New Results [Ada,Chattopadhyay,Cook,Fontes,P ‘12] (1)Using the fact that AvgPAR1 ≥ AvgPAR2, together with known IC lower bounds: Theorem AvgPAR2 of set intersection is Ω(n) (2) We prove strong tradeoffs for both worst-case PAR and avgPAR for Vickrey auctions. (3) Using compression [BBCR], it follows that any deterministic, low AvgPAR1 protocols can be compressed. Thus binary search protocol for millionaires implies a polylogn randomized protocol.

Important Open Questions IC=CC? IC in the multiparty NOF setting IC lower bounds for search problems Very important for proof complexity and circuit complexity Other applications of IC Data structures? Game Theory?

Thanks!