Lecture 18 The Main Coding Theory Problem (Section 4.7)

Slides:



Advertisements
Similar presentations
DCSP-8: Minimal length coding II, Hamming distance, Encryption Jianfeng Feng
Advertisements

Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 12 ERROR DETECTION & CORRECTION.
parity bit is 1: data should have an odd number of 1's
Error Correcting Codes Stanley Ziewacz 22M:151 Spring 2009.
Information Theory Introduction to Channel Coding Jalal Al Roumy.
1 Introduction to Computability Theory Lecture4: Non Regular Languages Prof. Amos Israeli.
1 Introduction to Computability Theory Lecture13: Mapping Reductions Prof. Amos Israeli.
Lecture 6: Huffman Code Thinh Nguyen Oregon State University.
1 Introduction to Computability Theory Lecture4: Non Regular Languages Prof. Amos Israeli.
Error detection and correction
Variable-Length Codes: Huffman Codes
Hamming Code Rachel Ah Chuen. Basic concepts Networks must be able to transfer data from one device to another with complete accuracy. Data can be corrupted.
exercise in the previous class (1)
Linear codes 1 CHAPTER 2: Linear codes ABSTRACT Most of the important codes are special types of so-called linear codes. Linear codes are of importance.
STATISTIC & INFORMATION THEORY (CSNB134)
Information Coding in noisy channel error protection:-- improve tolerance of errors error detection: --- indicate occurrence of errors. Source.
ENEE244-02xx Digital Logic Design Lecture 3. Announcements Homework 1 due next class (Thursday, September 11) First recitation quiz will be next Monday,
Introduction to Coding Theory. p2. Outline [1] Introduction [2] Basic assumptions [3] Correcting and detecting error patterns [4] Information rate [5]
Digital Communications I: Modulation and Coding Course Term Catharina Logothetis Lecture 12.
DIGITAL COMMUNICATIONS Linear Block Codes
Chapter 31 INTRODUCTION TO ALGEBRAIC CODING THEORY.
COMPSCI 102 Introduction to Discrete Mathematics.
Information Theory Linear Block Codes Jalal Al Roumy.
1 Coded modulation So far: Binary coding Binary modulation Will send R bits/symbol (spectral efficiency = R) Constant transmission rate: Requires bandwidth.
10.1 Chapter 10 Error Detection and Correction Data can be corrupted during transmission. Some applications require that errors be detected and.
Some Computation Problems in Coding Theory
Great Theoretical Ideas in Computer Science.
Turbo Codes. 2 A Need for Better Codes Designing a channel code is always a tradeoff between energy efficiency and bandwidth efficiency. Lower rate Codes.
Winter 2007CS244a Handout 141 CS244a: An Introduction to Computer Networks Handout 14: Error Detection and Correction Nick McKeown Professor of Electrical.
Section 3.2: Sequences and Summations. Def: A sequence is a function from a subset of the set of integers (usually the set of natural numbers) to a set.
Fidelity of a Quantum ARQ Protocol Alexei Ashikhmin Bell Labs  Classical Automatic Repeat Request (ARQ) Protocol  Quantum Automatic Repeat Request (ARQ)
INFORMATION THEORY Pui-chor Wong.
Hamming Distance & Hamming Code
Computer Architecture Error Correcting Codes Ralph Grishman Oct (Text pp and B-65-67) NYU.
Compression for Fixed-Width Memories Ori Rottenstriech, Amit Berman, Yuval Cassuto and Isaac Keslassy Technion, Israel.
1 Reliability-Based SD Decoding Not applicable to only graph-based codes May even help with some algebraic structure SD alternative to trellis decoding.
 2004 SDU Uniquely Decodable Code 1.Related Notions 2.Determining UDC 3.Kraft Inequality.
parity bit is 1: data should have an odd number of 1's
The Mathematics of Star Trek Workshop
Cyclic Codes 1. Definition Linear:
CSE322 PUMPING LEMMA FOR REGULAR SETS AND ITS APPLICATIONS
: An Introduction to Computer Networks
Set Topology MTH 251 Lecture # 8.
Induction and recursion
Proving the Correctness of Huffman’s Algorithm
Chapter 10 Error Detection And Correction
Welcome to the presentation. Linear Block Codes Almost all block codes used today belong to a subset called linear block codes. The exclusive OR of two.
EEL 3705 / 3705L Digital Logic Design
RS – Reed Solomon List Decoding.
Lecture 6 Instantaneous Codes and Kraft’s Theorem (Section 1.4)
EEC-484/584 Computer Networks
The sum of any two even integers is even.
Advanced Algorithms Analysis and Design
3.5 Minimum Cuts in Undirected Graphs
Block codes. encodes each message individually into a codeword n is fixed, Input/out belong to alphabet Q of cardinality q. The set of Q-ary n-tuples.
Information Redundancy Fault Tolerant Computing
CS 325: CS Hardware and Software Organization and Architecture
Error Detection and Correction
parity bit is 1: data should have an odd number of 1's
Huffman Coding Greedy Algorithm
Lecture 7 Information Sources; Average Codeword Length (Section 2.1)
Lecture 11 The Noiseless Coding Theorem (Section 3.4)
Proving the Correctness of Huffman’s Algorithm
Lecture 17 Making New Codes from Old Codes (Section 4.6)
Lecture 15 The Minimum Distance of a Code (Section 4.4)
Lecture 4 What are Codes? (Section 1.2)
Lecture 1 Introduction and preliminaries (Chapter 0)
Theory of Information Lecture 13
Lecture 8 Huffman Encoding (Section 2.2)
Presentation transcript:

Lecture 18 The Main Coding Theory Problem (Section 4.7) Theory of Information Lecture 13 Theory of Information Lecture 18 The Main Coding Theory Problem (Section 4.7)

Transmission Rate Efficiency --- large number of codewords for a given length Reliability --- (here) high minimum distance. Alas, these two main goals are at odds with each other! For easy comparison, we may assume that the source symbols are strings over the same alphabet as the codewords are. Then, roughly, each codeword of an r-ary code of size M captures a source “symbol” of length logrM. Thus the longer the codewords, the less the efficiency of encoding the source. Based on these remarks, a fair measurement of efficiency is what is called “transmission rate”: Definition The transmission rate R(C) of an r-ary (n,M)-code C is (logrM)/n. What are the transmission rates for the following codes? {0000,0001,0010,0011,0100,0101,0110,0111,1000,1001,1010,1011,1100,1101,1110,1111} {0000,0011,0101,0110,1001,1010,1100,1111} {0000,1111} {00,01,02,10,11,12,20,21,22} {00,11,22}

Error Correction Rate The longer the codewords, the higher the requirements on how many errors per transmitted codeword a good encoding scheme should allow to detect. Hence, a fair measurement of reliability is what is called “error correction rate”: Definition The error correction rate (C) of an (n,M,d)-code C is (d-1)/2 / n. What are the error correction rates for the following codes? {0000,0001,0010,0011,0100,0101,0110,0111,1000,1001,1010,1011,1100,1101,1110,1111} {0000,0011,0101,0110,1001,1010,1100,1111} {0000,1111} {00,01,02,10,11,12,20,21,22} {00,11,22} {11111,00000} {111111111,000000000} Give me a code with the error correction rate of 3 Give me a code with the error correction rate of 4 How’bout the transmission rates of the last four codes? We can see that the two sorts of rates are really at odds with each other.

The Famous Numbers Ar(n,d) For given values n (codeword length) and d (minimum distance), Ar(n,d) denotes the largest possible size M for which there exists an r-ary (n,M,d) code: Ar(n,d)=max{M | there exists an r-ary (n,M,d)-code} Any r-ary (n,M,d)-code C such that M=Ar(n,d) is called optimal. An optimal code thus has the highest possible transmission rate among the (n,d)-codes. Ar(n,d) is thus a very important number, yet very little is known so far about how to find it in a general case. Determining the values of Ar(n,d) has become known as the main coding theory problem. In some limited cases we can still succeed though.

Some Theorems on Ar(n,d) Theorem 4.7.1 A2(4,3)=2. PROOF. Let C be a (4,M,3)-code. By Lemma 4.6.1, we can assume that 0000 is in C. Since d(C)=3, any other codeword in C must be at distance at least 3 from 0000. This leaves us with 1110, 1101, 1011, 0111, 1111 as the only possible values for those other words. But no pair of these has distance 3 apart, and so only one can be included in C. Thus, C can have at most 2 codewords, implying Ar(4,3)2. And since C={0000,1110} is a (4,2,3)-code, we conclude Ar(4,3)=2. Theorem 4.7.4 1. Ar(n,d) rn for all 1dn 2. Ar(n,1)=rn 3. Ar(n,n)=r

Theory of Information Lecture 13 Homework 1. Exercise 19 of Section 4.7 2. What are the transmission rates for the following codes? a) {00011,11000,00110,01100,11100,11001,10011,00111} b) {000,001,002,110,111,112,220,221,222} 3. What are the error correction rates for the following codes? 4. We did not define error detection rate. You do it now (provide a formula).