Threshold Phenomena and Fountain Codes

Slides:



Advertisements
Similar presentations
Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen.
Advertisements

Jesper H. Sørensen, Toshiaki Koike-Akino, and Philip Orlik 2012 IEEE International Symposium on Information Theory Proceedings Rateless Feedback Codes.
Digital Fountains: Applications and Related Issues Michael Mitzenmacher.
Company LOGO F OUNTAIN C ODES, LT C ODES AND R APTOR C ODES Susmita Adhikari Eduard Mustafin Gökhan Gül.
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
Applied Algorithmics - week7
Error Control Code.
Digital Fountain Codes V. S
José Vieira Information Theory 2010 Information Theory MAP-Tele José Vieira IEETA Departamento de Electrónica, Telecomunicações e Informática Universidade.
Jump to first page A. Patwardhan, CSE Digital Fountains Main Ideas : n Distribution of bulk data n Reliable multicast, broadcast n Ideal digital.
D.J.C MacKay IEE Proceedings Communications, Vol. 152, No. 6, December 2005.
Capacity of Wireless Channels
1 Finite-Length Scaling and Error Floors Abdelaziz Amraoui Andrea Montanari Ruediger Urbanke Tom Richardson.
Cooperative Multiple Input Multiple Output Communication in Wireless Sensor Network: An Error Correcting Code approach using LDPC Code Goutham Kumar Kandukuri.
Codes for Deletion and Insertion Channels with Segmented Errors Zhenming Liu Michael Mitzenmacher Harvard University, School of Engineering and Applied.
Erasure Correcting Codes
Fountain Codes Amin Shokrollahi EPFL and Digital Fountain, Inc.
Sliding-Window Digital Fountain Codes for Streaming of Multimedia Contents Matta C.O. Bogino, Pasquale Cataldi, Marco Grangetto, Enrico Magli, Gabriella.
1 NETWORK CODING Anthony Ephremides University of Maryland - A NEW PARADIGM FOR NETWORKING - February 29, 2008 University of Minnesota.
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
Processing Along the Way: Forwarding vs. Coding Christina Fragouli Joint work with Emina Soljanin and Daniela Tuninetti.
RAPTOR CODES AMIN SHOKROLLAHI DF Digital Fountain Technical Report.
1 Verification Codes Michael Luby, Digital Fountain, Inc. Michael Mitzenmacher Harvard University and Digital Fountain, Inc.
The Role of Specialization in LDPC Codes Jeremy Thorpe Pizza Meeting Talk 2/12/03.
Accessing Multiple Mirror Sites in Parallel: Using Tornado Codes to Speed Up Downloads John Byers, Boston University Michael Luby, Digital Fountain, Inc.
Noise, Information Theory, and Entropy
Analysis of Iterative Decoding
Repairable Fountain Codes Megasthenis Asteris, Alexandros G. Dimakis IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 32, NO. 5, MAY /5/221.
Approximating the MST Weight in Sublinear Time Bernard Chazelle (Princeton) Ronitt Rubinfeld (NEC) Luca Trevisan (U.C. Berkeley)
Wireless Mobile Communication and Transmission Lab. Theory and Technology of Error Control Coding Chapter 7 Low Density Parity Check Codes.
Lecture 10: Error Control Coding I Chapter 8 – Coding and Error Control From: Wireless Communications and Networks by William Stallings, Prentice Hall,
Shifted Codes Sachin Agarwal Deutsch Telekom A.G., Laboratories Ernst-Reuter-Platz Berlin Germany Joint work with Andrew Hagedorn and Ari Trachtenberg.
An Optimal Partial Decoding Algorithm for Rateless Codes Valerio Bioglio, Rossano Gaeta, Marco Grangetto, and Matteo Sereno Dipartimento di Informatica.
Chih-Ming Chen, Student Member, IEEE, Ying-ping Chen, Member, IEEE, Tzu-Ching Shen, and John K. Zao, Senior Member, IEEE Evolutionary Computation (CEC),
X1X1 X2X2 Encoding : Bits are transmitting over 2 different independent channels.  Rn bits Correlation channel  (1-R)n bits Wireless channel Code Design:
Channel Capacity.
User Cooperation via Rateless Coding Mahyar Shirvanimoghaddam, Yonghui Li, and Branka Vucetic The University of Sydney, Australia IEEE GLOBECOM 2012 &
Threshold Phenomena and Fountain Codes Amin Shokrollahi EPFL Joint work with M. Luby, R. Karp, O. Etesami.
1 SNS COLLEGE OF ENGINEERING Department of Electronics and Communication Engineering Subject: Digital communication Sem: V Cyclic Codes.
MIMO continued and Error Correction Code. 2 by 2 MIMO Now consider we have two transmitting antennas and two receiving antennas. A simple scheme called.
Kai-Chao Yang VCLAB, NTHU 1.  Unequal Error Protection Rateless Codes for Scalable Information Delivery in Mobile Networks (INFOCOM 2007)  Rateless.
CprE 545 project proposal Long.  Introduction  Random linear code  LT-code  Application  Future work.
Andrew Liau, Shahram Yousefi, Senior Member, IEEE, and Il-Min Kim Senior Member, IEEE Binary Soliton-Like Rateless Coding for the Y-Network IEEE TRANSACTIONS.
Stochastic Networks Conference, June 19-24, Connections between network coding and stochastic network theory Bruce Hajek Abstract: Randomly generated.
Error Control Code. Widely used in many areas, like communications, DVD, data storage… In communications, because of noise, you can never be sure that.
DIGITAL COMMUNICATIONS Linear Block Codes
UEP LT Codes with Intermediate Feedback Jesper H. Sørensen, Petar Popovski, and Jan Østergaard Aalborg University, Denmark IEEE COMMUNICATIONS LETTERS,
1 Raptor codes for reliable multicast object delivery Michael Luby Digital Fountain.
Multi-Edge Framework for Unequal Error Protecting LT Codes H. V. Beltr˜ao Neto, W. Henkel, V. C. da Rocha Jr. Jacobs University Bremen, Germany IEEE ITW(Information.
Computer Science Division
Timo O. Korhonen, HUT Communication Laboratory 1 Convolutional encoding u Convolutional codes are applied in applications that require good performance.
Digital Communications I: Modulation and Coding Course Term Catharina Logothetis Lecture 9.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Raptor Codes Amin Shokrollahi EPFL. BEC(p 1 ) BEC(p 2 ) BEC(p 3 ) BEC(p 4 ) BEC(p 5 ) BEC(p 6 ) Communication on Multiple Unknown Channels.
OPTIMIZATION of GENERALIZED LT CODES for PROGRESSIVE IMAGE TRANSFER Suayb S. Arslan, Pamela C. Cosman and Laurence B. Milstein Department of Electrical.
Hongjie Zhu,Chao Zhang,Jianhua Lu Designing of Fountain Codes with Short Code-Length International Workshop on Signal Design and Its Applications in Communications,
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Hamming Distance & Hamming Code
Simulation of Finite Geometry LDPC code on the Packet Erasure channel Wu Yuchun July 2007 Huawei Hisi Company Ltd.
Lossless Compression-Statistical Model Lossless Compression One important to note about entropy is that, unlike the thermodynamic measure of entropy,
1 Implementation and performance evaluation of LT and Raptor codes for multimedia applications Pasquale Cataldi, Miquel Pedros Shatarski, Marco Grangetto,
Coding for Multipath TCP: Opportunities and Challenges Øyvind Ytrehus University of Bergen and Simula Res. Lab. NNUW-2, August 29, 2014.
8 Coding Theory Discrete Mathematics: A Concept-based Approach.
New Characterizations in Turnstile Streams with Applications
Chapter 5. Greedy Algorithms
Approximating the MST Weight in Sublinear Time
Chapter 6.
CRBcast: A Collaborative Rateless Scheme for Reliable and Energy-Efficient Broadcasting in Wireless Sensor/Actuator Networks Nazanin Rahnavard, Badri N.
Presentation transcript:

Threshold Phenomena and Fountain Codes Amin Shokrollahi EPFL Parts are joint work with M. Luby, R. Karp, O. Etesami

Communication on Multiple Unknown Channels BEC(p1) BEC(p2) BEC(p3) BEC(p6) BEC(p4) BEC(p5)

Example: Popular Download

Example: Peer-to-Peer

Example: Satellite

The erasure probabilities are unknown. Want to come arbitrarily close to capacity on each of the erasure channels, with minimum amount of feedback. Traditional codes don’t work in this setting since their rate is fixed. Need codes that can adapt automatically to the erasure rate of the channel.

What we Really Want Original content Encoded packets Users reconstruct Original content as soon as they receive enough packets Encoding Engine Transmission Want the encoded packets to be independently generated and unique Reconstruction time should depend only on size of content

Content Enc Digital buckets

Applications: Multi-site downloads Server 1 Server 2 Content Reception from multiple servers With the new coding technology one would avoid the coupon collector problem

Applications: Path Diversity Avoid congested network areas

Applications: Peer-2-Peer Sender 1 Sender 2 Sender 3 Rec 1 Rec2 Rec 3

Fountain Codes Sender sends a potentially limitless stream of encoded bits. Receivers collect bits until they are reasonably sure that they can recover the content from the received bits, and send STOP feedback to sender. Automatic adaptation: Receivers with larger loss rate need longer to receive the required information. Want that each receiver is able to recover from the minimum possible amount of received data, and do this efficiently.

Fountain Codes Fix distribution on , where is number of input symbols. For every output symbol sample independently from and add input symbols corresponding to sampled subset.

Fountain Codes Distribution on

Universality and Efficiency Want sequences of Fountain Codes for which the overhead is arbitrarily small [Efficiency] Want per-symbol-encoding to run in close to constant time, and decoding to run in time linear in number of output symbols.

Parameters Overhead is the fraction of extra symbols needed by the receiver as compared to the number of input symbols (i.e. if k(1+ε) is needed , overhead =ε) – want to be arbitrarily small Cost of encoding and decoding – linear in k Ratelessness – the number of output symbols is not determined apriori Space considerations (less important)

LT-Codes Invented by Michael Luby in 1998. First class of universal and almost efficient Fountain Codes Output distribution has a very simple form Encoding and decoding are very simple

LT-Codes LT-codes use a restricted distribution on : Fix distribution on Distribution is given by where is the Hamming weight of Parameters of the code are

The LT Coding Process Input symbols Weight table Choose 2 Random original symbols Choose weight XOR 2 Weight Prob 1 0.055 0.0004 0.3 2 0.1 3 0.08 4 100000 Weight table Insert header, and send

Decoding

Decoding

Decoding

Decoding

Decoding

Decoding

Decoding

Decoding

Decoding

Average Degree of Distribution should be

Average Degree of Distribution should be Not covered Prob. Non-coverage Prob. Decoding error Ω’(1) = average output node degree

Average Degree of Distribution should be Not covered Luby has designed universal LT-codes with average degree around and overhead i.e. the number of symbols needed is

So: Average degree constant means error probability constant How can we achieve constant workload per output symbol, and still guarantee vanishing error probability? Raptor codes achieve this!

Raptor Codes Use 2 phases: pre-code encodes k input symbols into intermediate code Apply LT to encode the intermediate code and transmit The LT-decoder only needs to recover (1-δ)n of the intermediate code w.h.p. This only requires constant degree!

Raptor Codes Traditional pre-code LT-light Input symbols d – fraction erasures LT-light Output symbols

Raptor Codes X Redundant Checks Not covered Redundant Checks If pre-code is chosen properly, then the LT-distribution can have constant average degree, leading to linear time encoding. Raptor Code is specified by the input length , precode and output distribution . How do we choose and ?

Special Raptor Codes: LT-Codes LT-Codes are Raptor Codes with trivial pre-code: Need average degree LT-Codes compensate for the lack of the pre-code with a rather intricate output distribution.

LT-lite Max degree Degree distribution: Average Degree: W.h.p can decode

What about pre-code? Can use regular LDPC codes such as Tornado, right-regular code, etc. But they have high average degree distributions…

Key Point An LDPC code can be systematic (I.e. input symbols appear in the set of the output symbols) Code rate (k/n) is arbitrarily close to 1, so most intermediate symbols have degree 1

Conclusions Encoding and Decoding can be done in time linear in k (or constant time per input symbol) The overhead is arbitrarily small as in regular LT codes Storage requirement is just the stretch factor n/k and k is arbitrarily close to n

Progressive Giant Component Analysis A different method for the analysis of the decoder: Want enough nodes of degree 2 so there exists a giant component in the induced (random) graph on input symbols.

Progressive Giant Component Analysis First giant component removes -fraction of input symbols. Residual distribution: Fraction of residual nodes of degree 2: Average degree of new induced graph: Condition: “Ideal distribution:”

Progressive Giant Component Analysis Analysis does not use “tree-assumption”, but only properties of induced graph. Analysis can be used to obtain error bounds for the decoding algorithm. It can also be used to obtain capacity-achieving distributions on the erasure channel. A modified version can be used to obtain for capacity-achieving distributions for other symmetric channels.

Nodes of Degree 2

Nodes of Degree 2 New output node of degree 2 Information Loss!

Fraction of Nodes of Degree 2 If there exists component of linear size (i.e., a giant component), then next output node of degree 2 has constant probability of being useless. Therefore, graph should not have giant component. This means that for capacity achieving degree distributions we must have: On the other hand, if then algorithm cannot start successfully. So, for capacity-achieving codes:

The -ary symmetric channel (large ) Double verification decoding (Luby-Mitzenmacher): If and are correct, then they verify . Remove all of them from graph and continue. Can be shown that number of correct output symbols needs to be at least Times number of input symbols.

The -ary symmetric channel (large ) More sophisticated algorithms: induced graph! If two input symbols are connected by a correct output symbol, and each of them is connected to a correct output symbol of degree one, then the input symbols are verified. Remove from them from graph.

The -ary symmetric channel (large ) More sophisticated algorithms: induced graph! More generally: if there is a path consisting of correct edges, and the two terminal nodes are connected to correct output symbols of degree one, then the input symbols get verified. (More complex algorithms.)

The -ary symmetric channel (large ) Limiting case: Giant component consisting of correct edges, two correct output symbols of degree one “poke” the component. So, ideal distribution “achieves” capacity.

Binary Memoryless Symmetric Channels What is the fraction of nodes of degree 2 for capacity-achieving Raptor Codes? where, in general and is the LLR of the channel.

General Symmetric Channels: Mimic Proof Proof is information theoretic: if fraction of nodes of degree 2 is larger by a constant, then : Expectation of the hyperbolic tangent of messages passed from input to output symbols at given round of BP is larger than a constant. This shows that So code cannot achieve capacity.

General Symmetric Channels: Mimic Proof Fraction of nodes of degree one for capacity-achieving Raptor Codes: Noisy observations of Therefore, if , and if denote output nodes of degree one, then So

Sequences Designed for the BEC Best designs (so far) Normalized SNR Eb/N0 0.000 0.067 0.135 0.194 0.267 0.331 0.391 0.459 0.522 0.584 0.650

Conclusions For LT- and Raptor codes, some decoding algorithms can be phrased directly in terms of subgraphs of graphs induced by output symbols of degree 2. This leads to a simpler analysis without the use the tree assumption. For the BEC, and for the q-ary symmetric channel (large q) we obtain essentially the same limiting capacity-achieving degree distribution, using the giant component analysis. An information theoretic analysis gives the optimal fraction of output nodes of degree 2 for general memoryless symmetric channels. A graph analysis reveals very good degree distributions, which perform very well experimentally.