STATISTIC & INFORMATION THEORY (CSNB134)

Slides:



Advertisements
Similar presentations
Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Advertisements

STATISTIC & INFORMATION THEORY (CSNB134) MODULE 12 ERROR DETECTION & CORRECTION.
Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
Information Theory EE322 Al-Sanie.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Random Variable A random variable X is a function that assign a real number, X(ζ), to each outcome ζ in the sample space of a random experiment. Domain.
Chapter 6 Information Theory
Middle Term Exam 03/04, in class. Project It is a team work No more than 2 people for each team Define a project of your own Otherwise, I will assign.
. Parameter Estimation and Relative Entropy Lecture #8 Background Readings: Chapters 3.3, 11.2 in the text book, Biological Sequence Analysis, Durbin et.
Fundamental limits in Information Theory Chapter 10 :
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
2015/6/15VLC 2006 PART 1 Introduction on Video Coding StandardsVLC 2006 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
Modeling OFDM Radio Channel Sachin Adlakha EE206A Spring 2001.
Entropy. Optimal Value Example The decimal number 563 costs 10  3 = 30 units. The binary number costs 2  10 = 20 units.  Same value as decimal.
Maximum likelihood (ML) and likelihood ratio (LR) test
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
3-1 Introduction Experiment Random Random experiment.
C4: DISCRETE RANDOM VARIABLES CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Longin Jan Latecki.
2015/7/12VLC 2008 PART 1 Introduction on Video Coding StandardsVLC 2008 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Albert Gatt Corpora and Statistical Methods. Probability distributions Part 2.
Some basic concepts of Information Theory and Entropy
Introduction to AEP In information theory, the asymptotic equipartition property (AEP) is the analog of the law of large numbers. This law states that.
Information and Coding Theory
2. Mathematical Foundations
Information Theory & Coding…
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Rev.S08 MAC 1140 Module 12 Introduction to Sequences, Counting, The Binomial Theorem, and Mathematical Induction.
PBG 650 Advanced Plant Breeding
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 8 INTRODUCTION TO INFORMATION THEORY.
Applied Business Forecasting and Regression Analysis Review lecture 2 Randomness and Probability.
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Channel Capacity.
Prepared by: Amit Degada Teaching Assistant, ECED, NIT Surat
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 7C PROBABILITY DISTRIBUTIONS FOR RANDOM VARIABLES ( NORMAL DISTRIBUTION)
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
Linawati Electrical Engineering Department Udayana University
Summer 2004CS 4953 The Hidden Art of Steganography A Brief Introduction to Information Theory  Information theory is a branch of science that deals with.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
DIGITAL COMMUNICATIONS Linear Block Codes
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Coding Theory Efficient and Reliable Transfer of Information
Prepared by: Engr. Jo-Ann C. Viñas 1 MODULE 2 ENTROPY.
Abdullah Aldahami ( ) April 6,  Huffman Coding is a simple algorithm that generates a set of variable sized codes with the minimum average.
Lecture 4: Lossless Compression(1) Hongli Luo Fall 2011.
CS654: Digital Image Analysis
. Parameter Estimation and Relative Entropy Lecture #8 Background Readings: Chapters 3.3, 11.2 in the text book, Biological Sequence Analysis, Durbin et.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Information theory (part II) From the form g(R) + g(S) = g(RS) one can expect that the function g( ) shall be a logarithm function. In a general format,
Entropy (YAC- Ch. 6)  Introduce the thermodynamic property called Entropy (S)  Entropy is defined using the Clausius inequality  Introduce the Increase.
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 11 COMPRESSION.
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Huffman Coding (2 nd Method). Huffman coding (2 nd Method)  The Huffman code is a source code. Here word length of the code word approaches the fundamental.
ENTROPY Entropy measures the uncertainty in a random experiment. Let X be a discrete random variable with range S X = { 1,2,3,... k} and pmf p k = P X.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Information Theory Information Suppose that we have the source alphabet of q symbols s 1, s 2,.., s q, each with its probability p(s i )=p i. How much.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
ECE 313 Probability with Engineering Applications Lecture 7
Introduction to Information theory
Corpora and Statistical Methods
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
A Brief Introduction to Information Theory
Sampling Theorems- Nyquist Theorem and Shannon-Hartley Theorem
Lecture 2: Basic Information Theory
Presentation transcript:

STATISTIC & INFORMATION THEORY (CSNB134) MODULE 9 INFORMATION CONTENT, ENTROPY & CODING EFFICIENCY

Information Uncertainty In Module 8, we have learned that the basic model of Information Theory is: Here, information is generated at the source, send through a channel and consume in the drain. In other words, information is being transmitted from the sender to the receiver. Prior to the transmission, the receiver has no idea what is the content of the information. This implies the concept of information as a random variable and ‘Information Uncertainty’ (i.e. the receiver is uncertain of the content of the information, until after he / she has received it down the transmission channel)

Information Uncertainty & Probability Consider the following statements: (a) Tomorrow, the sun will rise from the East. (b) The phone will ring in the next 1 hour. (c) It will snow in Malaysia this December. Everybody knows the sun always rises in the East. The probability of this event is almost 1. Thus, this statement hardly carries any information. The phone may or may not ring in the next 1 hour. The probability of this event is less that the probability of event in statement (a). Statement (b) carries more information than statement (a). It has never snowed in Malaysia. The probability of this event is almost 0. Statement (c) carries the most information amongst all.

Information Uncertainty & Probability (cont.) Therefore, we can conclude that the amount of information carries by each statement (or the information content of a single event) is inversely proportioned to the probability of that event. The formulae is: The unit of I(xi) is determined by the base of the logarithm. Since in the digital world, the basic unit is in bits, hence the formulae is:

Entropy Entropy is defined as the weighted sum of all information contents. Entropy is denoted as H(x) , and its formulae in base 2 is: Note: It is weighted / normalized by multiplying each information content with the probability of the event

Binary Entropy Function Consider a discrete binary source that emits a sequence of statistically independent symbols. The output is either ‘0’ with probability p or a ‘1’ with a probability 1-p. The entropy of the binary source is: The overall Binary Entropy Function can be plotted as: p

Exercise 1 Calculate the entropy of a binary source that has an equal probability of transmitting a ‘1’ and a ‘0’. Equal probability of transmitting a ‘1’ and a ‘0’ means p = 0.5 Note: If your calculator does not have the log2 function, then you may use this formulae: logbx = ln x / ln b

Exercise 2 Calculate the entropy of a binary source only transmit ‘1’ (i.e. transmit ‘1’ all the time!) Always transmitting a ‘1’ means p = 1

Exercise 3 Calculate the entropy of a binary source that transmit a ‘1’ after every 9 series of ‘0’. Transmit a ‘1’ after every 9 series of ‘0’ is similar to 0000000001 means p = 0.1 and 1-p = 0.9

Exercise 4 Assuming a binary source is used to cast the outcome of a fair dice. What is the entropy of the dice in bits? A fair dice always results in p = 1/6 for all possible 6 outcomes

Recaps… Entropy is the weighted sum of all information contents. In the example of the fair dice which outcome is represented in bits, the entropy is 2.584962. This can be concluded as : 2.584962 is the recommended average number of bits needed to sufficiently describe the outcome of a single cast. However, we know that by using fixed length coding we need to use an average of 3 bits to represent 6 symbols (Note: we denote Ṝ as the average bits / symbol, thus Ṝ = 3) . We know that 23 = 8, thus it is sub-optimal to represent only 6 symbols with 3 bits. Note: there is a difference between what is recommended and what is used in actual!

Average Number of Bits per Symbol Previously, we denote Ṝ as the average number of bits per symbol. In a fixed length coding (FLC), Ṝ is equal to n bits where 2n = maximum number of symbols than can be represented with n bits. In a fixed length coding, all symbols are represented by fixed same length number of bits. There is also another type of coding which is known as variable length coding (VLC). Here, all symbols need not necessarily be represented by a fixed same length number of bits. In both FLC and VLC the formulae for Ṝ is: where l(xi ) = length per symbol i

Coding Efficiency The formulae for entropy (H) is : Where as, the formulae for average number of bits / symbol (Ṝ) is: From here we can derive the coding efficiency (ἠ) as:

Exercise 5 Calculate the coding efficiency of FLC for the example of the fair dice where the entropy is 2.584962 and that the average number of bits / symbol using FLC is 3. The ideal optimal coding would yield efficiency of 1.

STATISTIC & INFORMATION THEORY (CSNB134) INFORMATION CONTENT, ENTROPY & CODING EFFICIENCY --END--