Collecting Correlated Information from a Sensor Network Micah Adler University of Massachusetts, Amherst.

Slides:



Advertisements
Similar presentations
1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Advertisements

Tight Bounds for Distributed Functional Monitoring David Woodruff IBM Almaden Qin Zhang Aarhus University MADALGO.
Optimal Space Lower Bounds for All Frequency Moments David Woodruff MIT
The Average Case Complexity of Counting Distinct Elements David Woodruff IBM Almaden.
Why Simple Hash Functions Work : Exploiting the Entropy in a Data Stream Michael Mitzenmacher Salil Vadhan And improvements with Kai-Min Chung.
Ulams Game and Universal Communications Using Feedback Ofer Shayevitz June 2006.
* Distributed Algorithms in Multi-channel Wireless Ad Hoc Networks under the SINR Model Dongxiao Yu Department of Computer Science The University of Hong.
Circuit and Communication Complexity. Karchmer – Wigderson Games Given The communication game G f : Alice getss.t. f(x)=1 Bob getss.t. f(y)=0 Goal: Find.
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
Applied Algorithmics - week7
Chapter 10 Shannon’s Theorem. Shannon’s Theorems First theorem:H(S) ≤ L n (S n )/n < H(S) + 1/n where L n is the length of a certain code. Second theorem:
Gillat Kol (IAS) joint work with Ran Raz (Weizmann + IAS) Interactive Channel Capacity.
Tight Bounds for Unconditional Authentication Protocols in the Moni Naor Gil Segev Adam Smith Weizmann Institute of Science Israel Modeland Shared KeyManual.
Onur G. Guleryuz & Ulas C.Kozat DoCoMo USA Labs, San Jose, CA 95110
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Collecting Correlated Information from Wireless Sensor Networks Presented by Bo Han Joint work with Aravind Srinivasan and Amol Deshpande.
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
SIA: Secure Information Aggregation in Sensor Networks Bartosz Przydatek, Dawn Song, Adrian Perrig Carnegie Mellon University Carl Hartung CSCI 7143: Secure.
Feb 25, 2003Mårten Trolin1 Previous lecture More on hash functions Digital signatures Message Authentication Codes Padding.
1 Staleness vs.Waiting time in Universal Discrete Broadcast Michael Langberg California Institute of Technology Joint work with Jehoshua Bruck and Alex.
Analysis of greedy active learning Sanjoy Dasgupta UC San Diego.
Mar 5, 2002Mårten Trolin1 Previous lecture More on hash functions Digital signatures Message Authentication Codes Padding.
Geographic Gossip: Efficient Aggregations for Sensor Networks Author: Alex Dimakis, Anand Sarwate, Martin Wainwright University: UC Berkeley Venue: IPSN.
Breaking the O(n 2 ) Bit Barrier: Scalable Byzantine Agreement with an Adaptive Adversary Valerie King Jared Saia Univ. of VictoriaUniv. of New Mexico.
Avraham Ben-Aroya (Tel Aviv University) Oded Regev (Tel Aviv University) Ronald de Wolf (CWI, Amsterdam) A Hypercontractive Inequality for Matrix-Valued.
A Hierarchical Energy-Efficient Framework for Data Aggregation in Wireless Sensor Networks IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, VOL. 55, NO. 3, MAY.
On the interdependence of routing and data compression in multi-hop sensor networks Anna Scaglione, Sergio D. Servetto.
Derandomizing LOGSPACE Based on a paper by Russell Impagliazo, Noam Nissan and Avi Wigderson Presented by Amir Rosenfeld.
EWSN 04 – Berlin, Jan. 20, 2004 Silence is Golden with High Probability: Maintaining a Connected Backbone in Wireless Sensor Networks Paolo Santi* Janos.
Mobility Increases Capacity In Ad-Hoc Wireless Networks Lecture 17 October 28, 2004 EENG 460a / CPSC 436 / ENAS 960 Networked Embedded Systems & Sensor.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
Scalable Information-Driven Sensor Querying and Routing for ad hoc Heterogeneous Sensor Networks Maurice Chu, Horst Haussecker and Feng Zhao Xerox Palo.
Lattices for Distributed Source Coding - Reconstruction of a Linear function of Jointly Gaussian Sources -D. Krithivasan and S. Sandeep Pradhan - University.
Compression with Side Information using Turbo Codes Anne Aaron and Bernd Girod Information Systems Laboratory Stanford University Data Compression Conference.
Linear Codes for Distributed Source Coding: Reconstruction of a Function of the Sources -D. Krithivasan and S. Sandeep Pradhan -University of Michigan,
Simulating independence: new constructions of Condensers, Ramsey Graphs, Dispersers and Extractors Boaz Barak Guy Kindler Ronen Shaltiel Benny Sudakov.
New Protocols for Remote File Synchronization Based on Erasure Codes Utku Irmak Svilen Mihaylov Torsten Suel Polytechnic University.
Genetic Regulatory Network Inference Russell Schwartz Department of Biological Sciences Carnegie Mellon University.
Decentralised load balancing in closed and open systems A. J. Ganesh University of Bristol Joint work with S. Lilienthal, D. Manjunath, A. Proutiere and.
When rate of interferer’s codebook small Does not place burden for destination to decode interference When rate of interferer’s codebook large Treating.
Streaming Algorithms Piotr Indyk MIT. Data Streams A data stream is a sequence of data that is too large to be stored in available memory Examples: –Network.
User Cooperation via Rateless Coding Mahyar Shirvanimoghaddam, Yonghui Li, and Branka Vucetic The University of Sydney, Australia IEEE GLOBECOM 2012 &
Threshold Phenomena and Fountain Codes Amin Shokrollahi EPFL Joint work with M. Luby, R. Karp, O. Etesami.
PODC Distributed Computation of the Mode Fabian Kuhn Thomas Locher ETH Zurich, Switzerland Stefan Schmid TU Munich, Germany TexPoint fonts used in.
Cooperative Communication in Sensor Networks: Relay Channels with Correlated Sources Brian Smith and Sriram Vishwanath University of Texas at Austin October.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Conjunctive Filter: Breaking the Entropy Barrier Daisuke Okanohara *1, *2 Yuichi Yoshida *1*3 *1 Preferred Infrastructure Inc. *2 Dept. of Computer Science,
Trade-offs Between Mobility and Density for Coverage in Wireless Sensor Networks Wei Wang, Vikram Srinivasan and Kee-Chaing Chua National University of.
Superposition encoding A distorted version of is is encoded into the inner codebook Receiver 2 decodes using received signal and its side information Decoding.
The Scaling Law of SNR-Monitoring in Dynamic Wireless Networks Soung Chang Liew Hongyi YaoXiaohang Li.
The Cost of Fault Tolerance in Multi-Party Communication Complexity Binbin Chen Advanced Digital Sciences Center Haifeng Yu National University of Singapore.
Umans Complexity Theory Lectures Lecture 7b: Randomization in Communication Complexity.
Artur Czumaj DIMAP DIMAP (Centre for Discrete Maths and it Applications) Computer Science & Department of Computer Science University of Warwick Testing.
Unique! coding for three different motivation flash codes network coding Slepian-Wolf coding test, take-home test.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Andrea CLEMENTI Radio Networks The Model Broadcast.
Error-Correcting Codes and Pseudorandom Projections Luca Trevisan U.C. Berkeley.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Samuel Cheng, Shuang Wang and Lijuan Cui University of Oklahoma
The Message Passing Communication Model David Woodruff IBM Almaden.
Jeffrey D. Ullman Stanford University.  A real story from CS341 data-mining project class.  Students involved did a wonderful job, got an “A.”  But.
Student: Omar Jarushi Instructor: Dr. Asaf Cohen 1.
Information Complexity Lower Bounds
Lecture 18: Uniformity Testing Monotonicity Testing
Cryptography Lecture 4.
Turnstile Streaming Algorithms Might as Well Be Linear Sketches
The Communication Complexity of Distributed Set-Joins
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 17 (2005) Richard Cleve DC 3524
Presentation transcript:

Collecting Correlated Information from a Sensor Network Micah Adler University of Massachusetts, Amherst

Fundamental Problem Collecting information from distributed sources. –Objective: correlations reduce bits that must be sent. Correlation examples in sensor networks: –Weather in geographic region. –Similar views of same image. Our focus: information theory –Number of bits sent. –Ignore network topology.

Modeling Correlation k sensor nodes each have n-bit string. Input drawn from distribution D. –Sample specifies all kn bits. –Captures correlations and a priori knowledge. Objective: –Inform server of all k strings. –Ideally: nodes send H(D) bits. –H(D): Binary entropy of D. x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 xkxk

Binary entropy of D Optimal code to describe sample from D: –Expected number of bits required ≈ H(D). Ranges from 0 to kn. –E asy if entire sample known by single node. Idea: shorter codewords for more likely samples. Challenge of our problem: input distributed.

Distributed Source Coding [Slepian-Wolf, 1973]: –Simultaneously encode r independent samples. –As r  , Bits sent by nodes  rH(D). Probability of error  0. Drawback: relies on r   –Recent research: try to remove this.

Recent Research on DSC Survey [XLC 2004]: over 50 recent DSC papers –All previous techniques: Significant restrictions on D and/or Large values of r required. –Can also be viewed as restriction on D. –Generalizing D most important open problem. Impossibility result: –There are D such that no encoding for small r achieves O(H(D)+k) bits sent from nodes. Our result: general D, r=1, O(H(D)+k) bits.

New approach Allow interactive communication! –Nodes receive “feedback” from server. Also utilized for DSC in [CPR 2003]. –Server at least as powerful as nodes. Power utilization: –Central issue for sensor networks. –Node sending: power intensive. –Node receiving requires less power. Analogy: portable radio vs. cellphone. x1x1 x2x2 x4x4 x5x5 xkxk x3x3

New approach Communication model: –Synchronous rounds: Nodes send bits to server. Server sends bits back to nodes. Nothing directly between nodes. Objectives: –Minimize bits sent by nodes. Ideally O(H(D)+k). –Minimize bits sent by server. –Minimize rounds of communication. x1x1 x2x2 x4x4 x5x5 xkxk x3x3

Asymmetric Commmunication Channels Adler-Maggs 1998: k=1 case. Subsequent work: [HL2002] [GGS2001] [W2000] [WAF2001] Other applications: –Circumventing web censorship [FBHBK2002] –Design of websites [BKLM2002] Sensor networks problem: natural parallelization. x D

Who knows what? Nodes: only know own string. –Can also assume they know distribution D. Server: knows distribution D. –Typical in work on DSC. –Some applications: D must be learned by server Most such cases: D varies with time. Crucial to have r as small as possible. D X1X1 X2X2 X3X3 X4X4 X5X5 X6X6 X7X7 X8X8, D

New Contributions New technique to communicate interactively: –O(H(D)+k) node bits. –O(kn + H(D) log n) server bits. –O(log min(k, H(D))) rounds. Lower bound: –kn bits must be exchanged if no error. If server is allowed error with probability ∆: –O(H(D)+k) node bits. –O(k log (kn/∆) + H(D) log n) server bits. –O(log min(k, H(D))) rounds.

General strategy Support uniform case: –D is uniform over set of possible inputs. General distributions: –Technique from [Adler-Maggs 1998] “Reduce” to support uniform case. –Requires modification to support uniform protocol. Allowing for Error: –Same protocol with some enhancements.

Support Uniform Input D: k-dimensional binary matrix –side length 2 n Choose X: uniform 1 entry of matrix. Server is given matrix, wants to know X. Node i given ith coordinate of X. H(D) = log M –M: number of possible inputs.

Basic Building Block Standard fingerprinting techniques: –Class of hash functions f: n-bit string  1 bit. For randomly chosen f, –If x  y, then Pr[ f(x) = f(y)] ≈ 1/2 Description of f requires O(log n) bits.

Not possible inputs Possible inputs Node 1 bits: Protocol for k=1 Server sends node log M fingerprint fcts. Node sends back resulting fingerprint bits. Intuition: each bit removes half inputs left.

What about k=2? Node 1 bits: Node 2 bits:

What about k=2? Node 1 bits: Node 2 bits:

First step: allow many rounds. Each round: –Server chooses one node. –That node sends single fingerprint bit. Objectives: –Ideal: each bit removes half remaining inputs. –Our goal: expected inputs removed is constant fraction. –Possibility: no node can send such a bit. Need to distinguish “useful” bits from “not useful” bits.

Balanced Bits Fingerprint bit sent by node i is balanced: –No value for i has > 1/2 possible inputs, –given all information considered so far. Balanced bits, in expectation, eliminate constant fraction of inputs. - Protocol goal: –Collect O(log M) balanced bits. –Don’t collect  (log M) unbalanced bits. Balanced bit:Unbalanced bit:

Objective: minimize rounds. Must send multiple bits from multiple nodes. –But shouldn’t send too many unbalanced bits! Difficulty: –Must decide how many at start of round. –As bits processed, matrix evolves. Node may only send unbalanced bits.

Number of bits sent per round Defined node: only one possible value left. –Should no longer send bits. First try: –Round i: k i undefined nodes, each send bits Possible problem: –Most nodes become defined at start of round i –Nodes might send total bits.

Protocol Description Phases: first round and second round. –First of phase i: undefined nodes send  b i  bits. Server processes bits in any order. Counts number of balanced bits. –Second: if node had any unbalanced bits Query if it has first heavy string. Continue until  (log M) balanced bits –Or until know entire input. –Send nodes remaining possibilities.

Performance of protocol Theorem: –Expected bits sent by nodes: O(k + log M) –Expected bits sent by server: O(kn + log M log n) –Expected number of rounds: O(min(log log M, log k))

Proof: O(log M + k) node bits. Key: if node i sends unbalanced bit in phase –Pr[i defined by end of phase]  1/2 Expected answers to heavy queries: O(1). Accounting for unbalanced bits: –Charge to bit sent by same node: In most recent balanced round –If none, then first round. Spread bits charged to round evenly.

Proof: O(log M + k) node bits. Expected unbalanced bits charged to any bit: Total balanced bits: O(log M) Total first round bits:

Proof: server bits and rounds Server bits: after O(log M) balanced bits –Pr[incorrect input not eliminated] = 1/M –Expected possible inputs remaining: O(1) Rounds: –Sparse phase: < 2/3 log M fingerprint bits –O(1) non-sparse phases. –O(min(log k, log log M)) sparse phases.

Conclusions New technique for collecting distributed and correlated information. Allows for arbitrary distributions and correlations. Single sample sufficient. Open problems: –Lower bound on rounds. –Achieving full Slepian-Wolf rate region. –Incorporating network topology.