Sensor Networks, Rate Distortion Codes, and Spin Glasses NTT Communication Science Laboratories Tatsuto Murayama In collaboration.

Slides:



Advertisements
Similar presentations
I NFORMATION CAUSALITY AND ITS TESTS FOR QUANTUM COMMUNICATIONS I- Ching Yu Host : Prof. Chi-Yee Cheung Collaborators: Prof. Feng-Li Lin (NTNU) Prof. Li-Yi.
Advertisements

Information Theory EE322 Al-Sanie.
Capacity of Wireless Channels
Onur G. Guleryuz & Ulas C.Kozat DoCoMo USA Labs, San Jose, CA 95110
Enhancing Secrecy With Channel Knowledge
Cooperative Multiple Input Multiple Output Communication in Wireless Sensor Network: An Error Correcting Code approach using LDPC Code Goutham Kumar Kandukuri.
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Chapter 6 Information Theory
On Systems with Limited Communication PhD Thesis Defense Jian Zou May 6, 2004.
Distributed Consensus with Limited Communication Data Rate Tao Li Key Laboratory of Systems & Control, AMSS, CAS, China ANU RSISE System and Control Group.
1 Data Persistence in Large-scale Sensor Networks with Decentralized Fountain Codes Yunfeng Lin, Ben Liang, Baochun Li INFOCOM 2007.
Digital Data Transmission ECE 457 Spring Information Representation Communication systems convert information into a form suitable for transmission.
Fundamental limits in Information Theory Chapter 10 :
2015/6/15VLC 2006 PART 1 Introduction on Video Coding StandardsVLC 2006 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
© 2005, it - instituto de telecomunicações. Todos os direitos reservados. Gerhard Maierbacher Scalable Coding Solutions for Wireless Sensor Networks IT.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
BASiCS Group University of California at Berkeley Generalized Coset Codes for Symmetric/Asymmetric Distributed Source Coding S. Sandeep Pradhan Kannan.
Comparative survey on non linear filtering methods : the quantization and the particle filtering approaches Afef SELLAMI Chang Young Kim.
Lattices for Distributed Source Coding - Reconstruction of a Linear function of Jointly Gaussian Sources -D. Krithivasan and S. Sandeep Pradhan - University.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 6-1 Chapter 6 The Normal Distribution and Other Continuous Distributions.
2015/7/12VLC 2008 PART 1 Introduction on Video Coding StandardsVLC 2008 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
Linear Codes for Distributed Source Coding: Reconstruction of a Function of the Sources -D. Krithivasan and S. Sandeep Pradhan -University of Michigan,
The Role of Specialization in LDPC Codes Jeremy Thorpe Pizza Meeting Talk 2/12/03.
Noise, Information Theory, and Entropy
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Noise, Information Theory, and Entropy
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
International Technology Alliance In Network & Information Sciences International Technology Alliance In Network & Information Sciences 1 Cooperative Wireless.
When rate of interferer’s codebook small Does not place burden for destination to decode interference When rate of interferer’s codebook large Treating.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
X1X1 X2X2 Encoding : Bits are transmitting over 2 different independent channels.  Rn bits Correlation channel  (1-R)n bits Wireless channel Code Design:
Channel Capacity.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Computer Vision – Compression(1) Hanyang University Jong-Il Park.
14 October, 2010LRI Seminar 2010 (Univ. Paris-Sud)1 Statistical performance analysis by loopy belief propagation in probabilistic image processing Kazuyuki.
DIGITAL COMMUNICATIONS Linear Block Codes
Coding Theory Efficient and Reliable Transfer of Information
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer.
Dr. Sudharman K. Jayaweera and Amila Kariyapperuma ECE Department University of New Mexico Ankur Sharma Department of ECE Indian Institute of Technology,
1 Central Limit Theorem The theorem states that the sum of a large number of independent observations from the same distribution has, under certain general.
Part 1: Overview of Low Density Parity Check(LDPC) codes.
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
Graduate School of Information Sciences, Tohoku University
Data Mining and Decision Support
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
CS294-9 :: Fall 2003 Joint Source/Channel Coding Ketan Mayer-Patel.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Image Processing Architecture, © Oleh TretiakPage 1Lecture 4 ECE-C490 Winter 2004 Image Processing Architecture Lecture 4, 1/20/2004 Principles.
A Low-Complexity Universal Architecture for Distributed Rate-Constrained Nonparametric Statistical Learning in Sensor Networks Avon Loy Fernandes, Maxim.
Raptor Codes Amin Shokrollahi EPFL. BEC(p 1 ) BEC(p 2 ) BEC(p 3 ) BEC(p 4 ) BEC(p 5 ) BEC(p 6 ) Communication on Multiple Unknown Channels.
INFORMATION THEORY Pui-chor Wong.
1 CSCD 433 Network Programming Fall 2013 Lecture 5a Digital Line Coding and other...
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Channel Coding: Part I Presentation II Irvanda Kurniadi V. ( ) Digital Communication 1.
DIGITAL COMMUNICATION. Introduction In a data communication system, the output of the data source is transmitted from one point to another. The rate of.
Institute for Experimental Mathematics Ellernstrasse Essen - Germany DATA COMMUNICATION introduction A.J. Han Vinck May 10, 2003.
1 CSCD 433 Network Programming Fall 2016 Lecture 4 Digital Line Coding and other...
The Viterbi Decoding Algorithm
Statistical-Mechanical Approach to Probabilistic Image Processing -- Loopy Belief Propagation and Advanced Mean-Field Method -- Kazuyuki Tanaka and Noriko.
Advanced Wireless Networks
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
2018/9/16 Distributed Source Coding Using Syndromes (DISCUS): Design and Construction S.Sandeep Pradhan, Kannan Ramchandran IEEE Transactions on Information.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Bayesian Deep Learning on a Quantum Computer
Presentation transcript:

Sensor Networks, Rate Distortion Codes, and Spin Glasses NTT Communication Science Laboratories Tatsuto Murayama In collaboration with Peter Davis March 7 th, 2008 at the Chinese Academy of Sciences

Problem Statement

Sensor Networks Sensor Sensors transmit their noisy observations independently. Computer Computer estimates the quantity of interest from sensor information. Network Network has a limited bandwidth constraint.

A Pessimistic Forecast Sensor Networks Central Unit Target Source Information loss via communications Information loss via sensing VS 《 Supply Side Economics 》 Semiconductors are going to be very small and also cheap, so they’d like to sell them a lot! Smartdusts, IC tags… Large-scale information integration Finite Network Capacity Finite Network Capacity Efficient use of the given bandwidth is required! Need a new information integration theory! High Noise Region High Noise Region Network is going to be large and dense! Network Capacity is limited

What to look for? Given a combined data rate, we examine the optimal aggregation level for sensor networks. Saturate Strategy (SS) Transmit as much sensor information as possible without data compression. Which strategy is outperforming? A small quantity of high quality statistics Large System Strategy (LSS) Transmit the overwhelming majority of compressed sensor information. A large quantity of low quality statistics

What to Evaluate? It is natural to introduce the following indicator function in decibel manner. Which Strategy is Outperforming to the Other? The large system strategy is outperforming when the indicator function is negative. The saturate strategy is outperforming when the indicator function is positive. The zero level corresponds to the strategic transition point if available.

What to Expect? Conjecture on the existence of the strategic transition point. Some Evidences At the low noise level, the indicator function should diverge to infinity. At the high noise level, the indicator function should converge to zero. Strategic Transition Point.

System Model

Target Information is a Bernoulli(1/2) Source. Environmental Noise is modeled by the Binary Symmetric Channel. Sensing Model Binary Symmetric Channel (BSC) The input alphabet is `flipped’ with a given probability. ObservationsSource

Communication Model To satisfy the bandwidth constraints, each sensor encodes its observation independently. CodewordsReproductions Nature of Bandwidth-Given Communication If the bandwidth is bigger than the entropy rate, revertible coding can be possible. If the bandwidth is smaller than the entropy rate, only non-revertible coding can be possible.

Estimation Model Collective estimation is done by applying the majority vote algorithm to the reproductions. Estimation Majority Vote Estimation is calculated from the reproductions by sequentially applying the following algorithm. In case of the `Ising’ alphabet

System Model Sensing Model Estimation Model Encoding Model Independent decoding process is forced Bitwise majority vote is concerned Assume purely random Source is observed

Case of Saturate Strategy Cost of comm.= # of sensors ( bits of info.) Moderate aggregation levels are possible. 2 messages saturate network. Encoding Decoding Estimation Sensing

Case of Large System Strategy Cost of comm. = # of sensors data rate We can make system as large as we want! Sensing Encoding Decoding Estimation Still 2 messages saturate network.

Collective Estimation

Rate Distortion Tradeoff Variety of communication reduces to a simple rate distortion tradeoff. Rate Distortion Tradeoff Each observation bit is flipped with the same probability. Black Box

Effective Distortion Under the stochastic description of the tradeoff, we introduce the effective distortion as follows. Then, our sensing and communications tasks reduces to a channel. The Channel Model The channel is labeled by effective distortion.

Formula for Finite Sensors Finite-scale Sensor Networks Given the number of sensors, we get with where

A Glimpse at Statistics In the large system limit, binomial distribution converges to normal distribution.

Changing Variables By the change of variables we have the following result.

Formula for Infinite Sensors Infinite-scale Sensor Networks Given only the noise and bandwidth, we get with where we naturally expect that

The Shannon Limit

《 Decoding 》 《 Encoding 》 Lossy Data Compression There exists tradeoff between compression rate and the resulting quality of reproduction. What is the best bound for the lossy compression? 《 Storage 》

Rate Distortion Theory Theory for compression beyond entropy rate. Best bound is the rate distortion function. Compression Rate Hamming Distortion ○ ×

Can the CEO be informed? Rate Distortion Function gives the best bound. Large System Strategy by optimal codes Leading Contribution Taylor Expansion Non-trivial regions are feasible Does LSS have any advantage over SS? The CEO can be informed!

Indicator Function In what condition the large system strategy outperforms the saturate strategy? Saturate Strategy is used as the `reference’ in the decibel measure. LSS SS Which is outperforming? LSS is outperforming when measure is negative. SS is outperforming when measure is positive.

Theoretical System Gain In the noisy environment, LSS is superior to SS! Existence of comparative advantage gives a strong motivation for making large systems.

Vector Quantization

Definition of VQ Any information bit belongs to the Voronoi region, and is replaced by its representative bit. Index map specifies the representative bits. Voronoi region is labeled by an index.

Gauge of Representative Bit Information is first divided into Voronoi regions, and then representative gauge is chosen.

Isolated Free Energy Free energy can be decoupled. Hamming Distortion can be derived. Isolated Model Reduces to Random Walk Statistics. Random Walk Statistics Cost Function ( Energy ) Exact Solution

Bit Error Probability Substitute exact solution into general formula. Theoretical Performance

Large System Gain Bit error probability in decibel measure Large system strategy is not so outperforming

Near Shannon Limit Algorithm

Rate Distortion Theory N bit sequence is encoded into M bit codeword. M bit codeword is decoded to reproduce N bit sequence, but not perfectly. Tradeoff relation between the rate R=M/N and the Hamming distortion D. Rate distortion function for random sequences

Sparse Matrix Coding Find a codeword sequence that satisfies: where the fidelity criterion: Boolean matrix A is characterized by K ones per row and C per column; an LDPC matrix. Bit wise reproduction errors are considered; the Hamming distortion measure D is selected.

Example: 4 bit sequence Set an LDPC matrix. Given a sequence: Find a codeword: Reproduce the original sequence.

Design Principle Algebraic constraints are represented in a graph. Probabilistic constraint is considered as a prior. Microscopic consistency might induce the macroscopic order of the frustrated system.

Low-resource Computation Introduce the mean field to avoid complex tasks. Eliminate many candidates of the solution by dynamical techniques. Hard Easy

TAP Approach A codeword bit is calculated by its marginal. Marginal probability is evaluated by heuristics.

Empirical Performance Message passing algorithm works very well.

Example of Saturate Strategy Six sensors transmit their original datawords. Sensing Transmission Estimation BER 9% 32.4k bps 5.4k bps BER 20.0%

Example of Large System Strategy Nine sensors transmit their codewords. Sensing Encoding & Transmission & Decoding Estimation BER 20.0% BER 24.7% BER 5% 5.4k bps 32.4k bps

Statistical Mechanics

Frustrated Free Energy Free energy cannot be decoupled. General formula for Hamming Distortion Frustrated model reduces to spin glass statistics. Saddle Point of Free Energy Cost Function ( Energy ) Approximation Replica Method

Bit Error Probability Substitute replica solution into general formula. Theoretical Performance Scaling Evaluation for Replica Solution

Characteristic Constant Constant: Saddle Point Equations Variance of order parameter: Non-negative entropy condition: Measure:

Large System Gain: K=2 Bit error probability in decibel measure Similar to the case of optimal random coding.

Large System Gain: K→∞ Bit error probability in decibel measure Coincides with optimal random coding.

Concluding Remarks We consider the problem of distributed sensing in a noisy environment. Limited bandwidth constraint induces tradeoff between reducing errors due to environmental noise and increasing errors due to lossy coding as number of sensors increases. Analysis shows threshold behavior for optimal number of sensors.

References

Analysis TM and M. Okada: `Rate Distortion Function in the Spin Glass State: A Toy Model’, Advances in Neural Information Processing Systems 15, , MIT Press (2003). Available at TM and P. Davis: `Rate Distortion Codes in Sensor Networks: A System-level Analysis’, Advances in Neural Information Processing Systems 18, , MIT Press (2006). Available at

Algorithms TM: `Statistical mechanics of the data compression theorem’, Journal of Physics A 35, L95-L100 (2002). Available at /35/8/101/a208l1.htmlhttp:// 4470/35/8/101/a208l1.html TM: `Thouless-Anderson-Palmer Approach for Lossy Compression’, Physical Review E 69, (R) (2004). Available at 05

Reviews TM and P. Davis: `Statistical mechanics of sensing and communications: Insights and techniques’, Journal of Physics: Conference Series 95, (2007). Available at /95/1http:// 6596/95/1 For more information, please google “tatsuto murayama” or “ 村山立人 ”.