The Zero-Error Capacity of Compound Channels Jayanth Nayak & Kenneth Rose University of California, Santa Barbara.

Slides:



Advertisements
Similar presentations
Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen.
Advertisements

DCSP-10 Jianfeng Feng Department of Computer Science Warwick Univ., UK
Graph-02.
Information Theory EE322 Al-Sanie.
B IPARTITE I NDEX C ODING Arash Saber Tehrani Alexandros G. Dimakis Michael J. Neely Department of Electrical Engineering University of Southern California.
The Cutoff Rate and Other Limits: Passing the Impassable Richard E. Blahut University of Illinois UIUC 5/4/20151.
Capacity of Wireless Channels
Submission May, 2000 Doc: IEEE / 086 Steven Gray, Nokia Slide Brief Overview of Information Theory and Channel Coding Steven D. Gray 1.
Chapter 6 Information Theory
Lossy Source Coding under a Maximum Distortion Constraint with Decoder Side- Information Jayanth Nayak 1, Ertem Tuncel 2, and Kenneth Rose 1 1 University.
Lihua Weng Dept. of EECS, Univ. of Michigan Error Exponent Regions for Multi-User Channels.
Zero-error source-channel coding with source side information at the decoder J. Nayak, E. Tuncel and K. Rose University of California, Santa Barbara.
Fundamental limits in Information Theory Chapter 10 :
Adaptive RLL Constrained Coding Seong Taek Chung Dec Stanford University.
Encoder and Decoder Optimization for Source-Channel Prediction in Error Resilient Video Transmission Hua Yang and Kenneth Rose Signal Compression Lab ECE.
Constrained Codes for PRML Panu Chaichanavong December 14, 2000 Partial Response Channel Maximum Likelihood Detection Constraints for PRML Examples Conclusion.
A Graph-based Framework for Transmission of Correlated Sources over Multiuser Channels Suhan Choi May 2006.
Improving the Performance of Turbo Codes by Repetition and Puncturing Youhan Kim March 4, 2005.
Compression with Side Information using Turbo Codes Anne Aaron and Bernd Girod Information Systems Laboratory Stanford University Data Compression Conference.
Distributed Video Coding Bernd Girod, Anne Margot Aaron, Shantanu Rane, and David Rebollo-Monedero IEEE Proceedings 2005.
Channel Polarization and Polar Codes
L/O/G/O FOUR CHANNEL WIRELESS TRANSMITTER AND RECEIVER.
Lecture 3 Outline Announcements: No class Wednesday Friday lecture (1/17) start at 12:50pm Review of Last Lecture Communication System Block Diagram Performance.
Analysis of Iterative Decoding
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
1 Channel Coding (II) Cyclic Codes and Convolutional Codes.
Information Coding in noisy channel error protection:-- improve tolerance of errors error detection: --- indicate occurrence of errors. Source.
(Important to algorithm analysis )
CODING/DECODING CONCEPTS AND BLOCK CODING. ERROR DETECTION CORRECTION Increase signal power Decrease signal power Reduce Diversity Retransmission Forward.
User Cooperation via Rateless Coding Mahyar Shirvanimoghaddam, Yonghui Li, and Branka Vucetic The University of Sydney, Australia IEEE GLOBECOM 2012 &
Error-Correction &Crosstalk Avoidance in DSM Busses Ketan Patel and Igor Markov University of Michigan Electrical Engineering & Computer Science 2003 ACM.
Prepared by: Amit Degada Teaching Assistant, ECED, NIT Surat
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
1/11/05 1 Capacity of Noiseless and Noisy Two-Dimensional Channels Paul H. Siegel Electrical and Computer Engineering Center for Magnetic Recording Research.
Communication Over Unknown Channels: A Personal Perspective of Over a Decade Research* Meir Feder Dept. of Electrical Engineering-Systems Tel-Aviv University.
Introduction to Coding Theory. p2. Outline [1] Introduction [2] Basic assumptions [3] Correcting and detecting error patterns [4] Information rate [5]
Communication Under Normed Uncertainties S. Z. Denic School of Information Technology and Engineering University of Ottawa, Ottawa, Canada C. D. Charalambous.
§2 Discrete memoryless channels and their capacity function
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
A Mathematical Theory of Communication Jin Woo Shin Sang Joon Kim Paper Review By C.E. Shannon.
Lecture 2 Outline Announcements: No class next Wednesday MF lectures (1/13,1/17) start at 12:50pm Review of Last Lecture Analog and Digital Signals Information.
DIGITAL COMMUNICATIONS Linear Block Codes
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Coding Theory Efficient and Reliable Transfer of Information
Transmission over composite channels with combined source-channel outage: Reza Mirghaderi and Andrea Goldsmith Work Summary STATUS QUO A subset Vo (with.
On Coding for Real-Time Streaming under Packet Erasures Derek Leong *#, Asma Qureshi *, and Tracey Ho * * California Institute of Technology, Pasadena,
Timo O. Korhonen, HUT Communication Laboratory 1 Convolutional encoding u Convolutional codes are applied in applications that require good performance.
Digital Communications I: Modulation and Coding Course Term Catharina Logothetis Lecture 9.
Interactive Channel Capacity. [Shannon 48]: A Mathematical Theory of Communication An exact formula for the channel capacity of any noisy channel.
Presented by Minkoo Seo March, 2006
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Source Encoder Channel Encoder Noisy channel Source Decoder Channel Decoder Figure 1.1. A communication system: source and channel coding.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Error-Correcting Code
Bus Encoding to Prevent Crosstalk Delay Bert Victor and Kurt Keutzer ICCAD 2001.
Information Theory & Coding for Digital Communications Prof JA Ritcey EE 417 Source; Anderson Digital Transmission Engineering 2005.
Network Topology Single-level Diversity Coding System (DCS) An information source is encoded by a number of encoders. There are a number of decoders, each.
Channel Coding: Part I Presentation II Irvanda Kurniadi V. ( ) Digital Communication 1.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
A Graph Theoretic Approach to Cache-Conscious Placement of Data for Direct Mapped Caches Mirza Beg and Peter van Beek University of Waterloo June
Chapter 4: Information Theory. Learning Objectives LO 4.1 – Understand discrete and continuous messages, message sources, amount of information and its.
Ch4. Zero-Error Data Compression Yuan Luo. Content  Ch4. Zero-Error Data Compression  4.1 The Entropy Bound  4.2 Prefix Codes  Definition and.
The Viterbi Decoding Algorithm
Advanced Wireless Networks
2018/9/16 Distributed Source Coding Using Syndromes (DISCUS): Design and Construction S.Sandeep Pradhan, Kannan Ramchandran IEEE Transactions on Information.
S Digital Communication Systems
Chapter 6.
Homework #2 Due May 29 , Consider a (2,1,4) convolutional code with g(1) = 1+ D2, g(2) = 1+ D + D2 + D3 a. Draw the.
Lihua Weng Dept. of EECS, Univ. of Michigan
Presentation transcript:

The Zero-Error Capacity of Compound Channels Jayanth Nayak & Kenneth Rose University of California, Santa Barbara

2 Outline Preliminaries Problem definition Capacity –Neither side has side-information –Encoder has side-information –Decoder has side-information Comparison with vanishing-error case

3 Zero-Error vs. Vanishing-Error Vanishing-error Case – tends to zero as block length tends to infinity. Zero-error case – zero at all block lengths –Tools: graph theory and combinatorics

4 Discrete Memoryless Channel Characteristic graph of the channel Example Channel transition probability   a b d e c GXGX Undirected

5 Channel Code Code  symbols pair-wise connected = clique Example (cont.) –Scalar code: 1-use capacity = – is the clique number bits GXGX

6 Channel Code uses of the channel –Fan-out set is Cartesian product of fan-out sets at each coordinate – = size of largest pairwise connected set Connected

7 Channel Code Zero error capacity (Shannon ’56) Depends only on characteristic graph – Shannon capacity of a graph bits/channel use

8 Other Graph Capacities Shannon capacity of a set of graphs: –Asymptotic size of set of sequences connected with respect to every graph from –Defined by Cohen et al. ‘88 Directed graphs Connected

9 Other Graph Capacities Sperner capacity of a graph – largest pairwise connected set : connected with respect to every graph Sperner capacities defined by Gargano et al. ’94 –Motivated by extremal set theory

10 The Compound Channel Once chosen, DMC remains constant throughout the transmission. EncoderDecoder Side Information = S S

11 Capacity Conventional Case: –Dobrushin ‘59, Wolfowitz ‘60, Breiman ‘59 Zero-Error Case: –Cohen et al. ‘88, Gargano et al. ‘94 –Capacity expression accurate only when decoder is informed.

12 Characteristic Set of Graphs Fan-out set – –Vertex set – –Directed graph

13 Capacity of Compound Channel Every code is a pairwise connected set with respect to and vice versa

14 Encoder Informed of S If >0, encoder sends side- information to decoder. Needs constant number of channel uses. = 0 case: – contains an edge-free graph

15 Decoder Informed of S Decoder can change with channel – Therefore –

16 Conventional vs. Zero-Error Conventional Capacities Zero-Error Capacities Reason: Decoder can identify channel with high probability

Thank you