UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.

Slides:



Advertisements
Similar presentations
Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen.
Advertisements

CY2G2 Information Theory 1
Lecture 4 (week 2) Source Coding and Compression
Sampling and Pulse Code Modulation
Arithmetic Coding. Gabriele Monfardini - Corso di Basi di Dati Multimediali a.a How we can do better than Huffman? - I As we have seen, the.
Information Theory EE322 Al-Sanie.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
SIMS-201 Compressing Information. 2  Overview Chapter 7: Compression Introduction Entropy Huffman coding Universal coding.
Chain Rules for Entropy
Chapter 6 Information Theory
Lecture04 Data Compression.
Fundamental limits in Information Theory Chapter 10 :
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Data Structures – LECTURE 10 Huffman coding
Information Theory Rong Jin. Outline  Information  Entropy  Mutual information  Noisy channel model.
Variable-Length Codes: Huffman Codes
Fundamentals of Multimedia Chapter 7 Lossless Compression Algorithms Ze-Nian Li and Mark S. Drew 건국대학교 인터넷미디어공학부 임 창 훈.
CSI Uncertainty in A.I. Lecture 201 Basic Information Theory Review Measuring the uncertainty of an event Measuring the uncertainty in a probability.
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Source Coding Hafiz Malik Dept. of Electrical & Computer Engineering The University of Michigan-Dearborn
Noise, Information Theory, and Entropy
1 Lossless Compression Multimedia Systems (Module 2) r Lesson 1: m Minimum Redundancy Coding based on Information Theory: Shannon-Fano Coding Huffman Coding.
Noise, Information Theory, and Entropy
Basics of Compression Goals: to understand how image/audio/video signals are compressed to save storage and increase transmission efficiency to understand.
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Huffman Coding Vida Movahedi October Contents A simple example Definitions Huffman Coding Algorithm Image Compression.
§1 Entropy and mutual information
Information Theory & Coding…
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Fundamentals of Digital Communication 2 Digital communication system Low Pass Filter SamplerQuantizer Channel Encoder Line Encoder Pulse Shaping Filters.
Channel Coding Part 1: Block Coding
§4 Continuous source and Gaussian channel
Channel Capacity
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Channel Capacity.
§3 Discrete memoryless sources and their rate-distortion function §3.1 Source coding §3.2 Distortionless source coding theorem §3.3 The rate-distortion.
Information Theory and Coding System EMCS 676 Fall 2014 Prof. Dr. Md. Imdadul Islam
Prepared by: Amit Degada Teaching Assistant, ECED, NIT Surat
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
1 Information in Continuous Signals f(t) t 0 In practice, many signals are essentially analogue i.e. continuous. e.g. speech signal from microphone, radio.
Lossless Compression CIS 465 Multimedia. Compression Compression: the process of coding that will effectively reduce the total number of bits needed to.
§2 Discrete memoryless channels and their capacity function
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
DIGITAL COMMUNICATIONS Linear Block Codes
Huffman Code and Data Decomposition Pranav Shah CS157B.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Coding Theory Efficient and Reliable Transfer of Information
Additive White Gaussian Noise
Source Coding Efficient Data Representation A.J. Han Vinck.
Abdullah Aldahami ( ) April 6,  Huffman Coding is a simple algorithm that generates a set of variable sized codes with the minimum average.
Lecture 4: Lossless Compression(1) Hongli Luo Fall 2011.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
INFORMATION THEORY Pui-chor Wong.
1 Data Compression Hae-sun Jung CS146 Dr. Sin-Min Lee Spring 2004.
Chapter 7 Lossless Compression Algorithms 7.1 Introduction 7.2 Basics of Information Theory 7.3 Run-Length Coding 7.4 Variable-Length Coding (VLC) 7.5.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Huffman Coding (2 nd Method). Huffman coding (2 nd Method)  The Huffman code is a source code. Here word length of the code word approaches the fundamental.
UNIT –V INFORMATION THEORY EC6402 : Communication TheoryIV Semester - ECE Prepared by: S.P.SIVAGNANA SUBRAMANIAN, Assistant Professor, Dept. of ECE, Sri.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Chapter 4: Information Theory. Learning Objectives LO 4.1 – Understand discrete and continuous messages, message sources, amount of information and its.
Information Theory Information Suppose that we have the source alphabet of q symbols s 1, s 2,.., s q, each with its probability p(s i )=p i. How much.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Introduction to Information theory
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
Subject Name: Information Theory Coding Subject Code: 10EC55
Distributed Compression For Binary Symetric Channels
CSE 589 Applied Algorithms Spring 1999
Presentation transcript:

UNIT I

Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity below which a signal cannot be compressed. Capacity is the ultimate transmission rate for reliable communication over noisy channel. Capacity is the ultimate transmission rate for reliable communication over noisy channel. Entropy: The probabilistic behavior of a source of information. Entropy: The probabilistic behavior of a source of information. K-1 K-1 H(X) = ∑ pk log2(1/pk) H(X) = ∑ pk log2(1/pk) k=0 k=0 Capacity: Intrinsic ability of a channel to convey information. Capacity: Intrinsic ability of a channel to convey information. Information: Measure of uncertainty of an event. Information: Measure of uncertainty of an event. I k = log2(1/pk)

Source coding Theorem Given a discrete memoryless source of entropy H(X). The average length of code word L for any distortionless source-encoding scheme is bounded as L >= H(X). Given a discrete memoryless source of entropy H(X). The average length of code word L for any distortionless source-encoding scheme is bounded as L >= H(X). According to the theorem, the entropy H(X) represents a fundamental limit on L. Therefore, According to the theorem, the entropy H(X) represents a fundamental limit on L. Therefore, Lmin = H(X).

Huffman coding Definition: A minimal variable-length character coding based on the frequency of each character. Definition: A minimal variable-length character coding based on the frequency of each character. The Huffman coding algorithm proceeds as follows The Huffman coding algorithm proceeds as follows The source symbols are listed in order of decreasing probability. The two source symbols of lowest probability are assigned a0 and a1. The source symbols are listed in order of decreasing probability. The two source symbols of lowest probability are assigned a0 and a1. These two symbols are regarded as being combined into a new source symbol with probability equal to the sum of the two original probabilities. These two symbols are regarded as being combined into a new source symbol with probability equal to the sum of the two original probabilities. The procedure is repeated until we are left with final list of source statistics on only two for which a0 and a1 are assigned. The procedure is repeated until we are left with final list of source statistics on only two for which a0 and a1 are assigned.

Huffman coding An example of the Huffman encoding algorithm.

Shannon Fano coding Definition: A variable-length coding based on the frequency of occurrence of each character. Divide the characters into two sets with the frequency of each set as close to half as possible, and assign the sets either 0 or 1 coding. Repeatedly divide the sets until each character has a unique coding. Definition: A variable-length coding based on the frequency of occurrence of each character. Divide the characters into two sets with the frequency of each set as close to half as possible, and assign the sets either 0 or 1 coding. Repeatedly divide the sets until each character has a unique coding.sets Note: Shannon-Fano is a minimal prefix code. Huffman is optimal for character coding (one character-one code word) and simple to program. Arithmetic coding is better still, since it can allocate fractional bits, but is more complicated and has patents. Note: Shannon-Fano is a minimal prefix code. Huffman is optimal for character coding (one character-one code word) and simple to program. Arithmetic coding is better still, since it can allocate fractional bits, but is more complicated and has patents.

Discrete Memory less channels When the probabilities of selection of successive events are independent, then the source is said to be discrete memoryless source or zero memory source.

Channel capacity Channel capacity of a discrete memoryless channel is the maximum mutual information I(X,Y) in any single use of channel, where the maximization is overall possible input probability distributions { p(x j)} on x. C = max I(X,Y)

Channel coding Theorem Let a discrete memoryless source with an alphabet x have entropy H(X) and produces symbols every Ts seconds. Let a discrete memory less channel have capacity C and be used every Tc seconds, then if there exist a coding scheme for which the source information can be transmitted over the channel and be reconstructed at the receiver with very small probability of error.

Channel capacity Theorem Shannon’s information capacity theorem states that the channel capacity of a continuous channel of bandwidth W Hz, perturbed by bandlimited Gaussian noise of power spectral density n0 /2, is given by Cc = W log2(1 + SN) bits/s (32.1) where S is the average transmitted signal power and the average noise power is where S is the average transmitted signal power and the average noise power is W N =∫ n0/2 dw = n0W (32.2) N =∫ n0/2 dw = n0W (32.2) -W -W