The Channel and Mutual Information

Slides:



Advertisements
Similar presentations
Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen.
Advertisements

Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Chapter 8 Channel Capacity. bits of useful info per bits actually sent Change in entropy going through the channel (drop in uncertainty): average uncertainty.
Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Binary Symmetric channel (BSC) is idealised model used for noisy channel. symmetric p( 01) =p(10)
Sampling and Pulse Code Modulation
Chapter 10 Shannon’s Theorem. Shannon’s Theorems First theorem:H(S) ≤ L n (S n )/n < H(S) + 1/n where L n is the length of a certain code. Second theorem:
Information Theory EE322 Al-Sanie.
Chain Rules for Entropy
Chapter 6 Information Theory
. Parameter Estimation and Relative Entropy Lecture #8 Background Readings: Chapters 3.3, 11.2 in the text book, Biological Sequence Analysis, Durbin et.
The Gibbs sampler Suppose f is a function from S d to S. We generate a Markov chain by consecutively drawing from (called the full conditionals). The n’th.
Fundamental limits in Information Theory Chapter 10 :
DATA ANALYSIS Module Code: CA660 Lecture Block 2.
資訊理論 授課老師 : 陳建源 研究室 : 法 401 網站 Ch4: Channel.
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Information Theory Rong Jin. Outline  Information  Entropy  Mutual information  Noisy channel model.
資訊理論 授課老師 : 陳建源 研究室 : 法 401 網站 Ch2: Basic Concepts.
資料壓縮 授課老師 : 陳建源 研究室 : 法 401 網站 資訊理論基本概念.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
Information Theory and Security
X= {x 0, x 1,….,x J-1 } Y= {y 0, y 1, ….,y K-1 } Channel Finite set of input (X= {x 0, x 1,….,x J-1 }), and output (Y= {y 0, y 1,….,y K-1 }) alphabet.
Review of Probability Theory. © Tallal Elshabrawy 2 Review of Probability Theory Experiments, Sample Spaces and Events Axioms of Probability Conditional.
1 Statistical NLP: Lecture 5 Mathematical Foundations II: Information Theory.
§1 Entropy and mutual information
2. Mathematical Foundations
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
CY2G2 Information Theory 5
Exercise in the previous class p: the probability that symbols are delivered correctly C: 1 00 → → → → What is the threshold.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Channel Capacity.
1.Check if channel capacity* can cope with source information rate, if yes, source coding can proceed. Understand why. (i) Calculate Source entropy from.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
2-1 Sample Spaces and Events Random Experiments Figure 2-1 Continuous iteration between model and physical system.
2-1 Sample Spaces and Events Random Experiments Figure 2-1 Continuous iteration between model and physical system.
Chapter 4 Probability ©. Sample Space sample space.S The possible outcomes of a random experiment are called the basic outcomes, and the set of all basic.
§2 Discrete memoryless channels and their capacity function
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
DIGITAL COMMUNICATIONS Linear Block Codes
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Coding Theory Efficient and Reliable Transfer of Information
CHAPTER 5 SIGNAL SPACE ANALYSIS
1 Value of information – SITEX Data analysis Shubha Kadambe (310) Information Sciences Laboratory HRL Labs 3011 Malibu Canyon.
12/7/20151 Math b Conditional Probability, Independency, Bayes Theorem.
CSCI 171 Presentation 9 Matrix Theory. Matrix – Rectangular array –i th row, j th column, i,j element –Square matrix, diagonal –Diagonal matrix –Equality.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
2IS80 Fundamentals of Informatics
Digital Image Processing Lecture 22: Image Compression
11.7 Continued Probability. Independent Events ► Two events are independent if the occurrence of one has no effect on the occurrence of the other ► Probability.
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Mutual Information, Joint Entropy & Conditional Entropy
Mutual Information and Channel Capacity Multimedia Security.
Channel Coding: Part I Presentation II Irvanda Kurniadi V. ( ) Digital Communication 1.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Chapter 4: Information Theory. Learning Objectives LO 4.1 – Understand discrete and continuous messages, message sources, amount of information and its.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Digital Multimedia Coding
Chapter 6.
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
Subject Name: Information Theory Coding Subject Code: 10EC55
Problem 1.
Lecture 2: Basic Information Theory
Lecture 9 Entropy (Section 3.1, a bit of 3.2)
Theory of Information Lecture 13
Presentation transcript:

The Channel and Mutual Information Chapter 7 The Channel and Mutual Information

Information Through a Channel symbols can’t be swallowed A a1 : : aq b1 : : bs B alphabet of symbols sent alphabet of symbols received P(bj|ai) or randomly generated For example, in an error correcting code over a noisy channel, s ≥ q. If two symbols sent are indistinguishable when received, s < q. Characterize a stationary channel by a matrix of conditional probabilities: Compare: noise (randomness) versus distortion (permutation) received row column sent s Pi,j Pi,j = P(bj | ai) P = q 7.1, 7.2, 7.3

[p(a1) … p(aq)]P = [p(b1) … p(bs)] For p(ai) = probability of source symbols, let p(bj) = probability of being received [p(a1) … p(aq)]P = [p(b1) … p(bs)] no noise: Pi,j = I; p(bj) = p(aj) all noise: Pi,j = 1/s; p(bj) = 1/s The probability that ai was sent and bj was received is: Baye’s Theorem Intuition: given b_j, some a_i must have been sent P(ai, bj) = p(ai) ∙ P(bj | ai) = p(bj) ∙ P(ai | bj). [coincidental probability] So if p(bj) ≠ 0, the backwards conditional probabilities are: 7.1, 7.2, 7.3

Binary symmetric Channel P0,0 a = 0 b = 0 p(a = 0) = p p(a = 1) = 1 − p P0,0 = P1,1 = P P0,1 = P1,0 = Q P0,1 P1,0 P1,1 a = 1 b = 1 a = 0 a = 1 p(b = 0) p(b = 1) P Q Q P ( pP + (1 − p)Q pQ + (1 − p)P ) (p 1−p) = 7.4

Backwards conditional probabilities No noise All noise P(a = 0 | b = 0) = Pp 1 p Pp + Q(1−p) P(a = 1 | b = 0) Q(1−p) 1 − p P(a = 0 | b = 1) Qp Qp + P(1−p) P(a = 1 | b = 1) P(1−p) Is this the only situation where the forwards and backwards matrix is the same? P = 1 Q = 0 P = Q = ½ If p = 1 − p = ½ (equiprobable) then: P(a = 1 | b = 0) = P(a = 0 | b = 1) = Q P(a = 0 | b = 0) = P(a = 1 | b = 1) = P P Q Q P 7.4

System Entropies H(A) Input entropy H(B) Output entropy H(A| B) H(B| A) H(A) Input entropy H(B) Output entropy condition on bj average over all bj Similarly The information loss in the channel, called equivocation (or noise entropy). The average uncertainty about the symbol sent or received. 7.5

H(A, B) = H(A, B) Joint Entropy Define : H(B | A) H(A) H(A | B) H(B) Intuition: taking snapshots A B Define : H(A| B) H(B| A) H(B | A) H(A) The signal H(A) is the information you want. The signal and noise H(B) is the information you have. H(B|A) is the information you don’t want (from noise in the channel). H(A | B) H(A, B) = H(B) 7.5

a priori a posteriori H(A, B) joint Mutual Information P(ai | bj) H(B| A) P(ai | bj) p(ai) I(A; B) The amount of information they are sharing corresponds to Information gain upon receiving bj : I(ai) − I(ai | bj) . shared I(A; B) is the reduction in the uncertainty of A due to the knowledge of B. By symmetry: If ai and bj are independent (all noise), then P(ai , bj) = p(ai) ∙ p(bj) and hence P(ai | bj) = p(ai)  I(ai ; bj) = 0. No information gained in channel. 7.6

= H(A) + H(B) − H(A, B) ≥ 0  H(A, B)  H(A) + H(B) Average over all ai: Similarly: from symmetry By Gibbs I(A; B) ≥ 0. Equality only if P(ai, bj) = p(ai)∙p(bj) [independence]. Can derive these inequalities from the figure on the previous slide. = H(A) + H(B) − H(A, B) ≥ 0  H(A, B)  H(A) + H(B) We know H(A, B) = H(A) + H(B | A) = H(B) + H(A | B).  I(A ; B) = H(A) − H(A | B) = H(B) − H(B | A) ≥ 0  H(A | B)  H(A) and H(B | A)  H(B). 7.6