©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.

Slides:



Advertisements
Similar presentations
Sampling and Pulse Code Modulation
Advertisements

Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
Random Variables ECE460 Spring, 2012.
Arithmetic Coding. Gabriele Monfardini - Corso di Basi di Dati Multimediali a.a How we can do better than Huffman? - I As we have seen, the.
Information Theory EE322 Al-Sanie.
Quantization Prof. Siripong Potisuk.
Chapter 6 Information Theory
SWE 423: Multimedia Systems
Fundamental limits in Information Theory Chapter 10 :
Point estimation, interval estimation
Spatial and Temporal Data Mining
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Evaluating Hypotheses
Losslessy Compression of Multimedia Data Hao Jiang Computer Science Department Sept. 25, 2007.
Chapter 11 Multiple Regression.
Continuous Random Variables and Probability Distributions
Experimental Evaluation
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
Information Theory and Security
Noise, Information Theory, and Entropy
Standard error of estimate & Confidence interval.
Noise, Information Theory, and Entropy
Data Sampling & Nyquist Theorem Richa Sharma Dept. of Physics And Astrophysics University of Delhi.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Formatting and Baseband Modulation
Probability Theory and Random Processes
Chapter 7 Estimation: Single Population
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
Channel Capacity.
Image Compression (Chapter 8) CSC 446 Lecturer: Nada ALZaben.
Image Compression – Fundamentals and Lossless Compression Techniques
Image Compression Fasih ur Rehman. Goal of Compression Reduce the amount of data required to represent a given quantity of information Reduce relative.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Chapter 17 Image Compression 17.1 Introduction Redundant and irrelevant information  “Your wife, Helen, will meet you at Logan Airport in Boston.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
1 Quantization Error Analysis Author: Anil Pothireddy 12/10/ /10/2002.
Introduction to Digital Signals
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Source Coding Efficient Data Representation A.J. Han Vinck.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Lecture 4: Lossless Compression(1) Hongli Luo Fall 2011.
CS654: Digital Image Analysis
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Digital Image Processing Lecture 22: Image Compression
Discrete-time Random Signals
Continuous Random Variables and Probability Distributions
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 11 COMPRESSION.
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
ECE 101 An Introduction to Information Technology Information Coding.
Fundamentals of Multimedia Chapter 6 Basics of Digital Audio Ze-Nian Li and Mark S. Drew 건국대학교 인터넷미디어공학부 임 창 훈.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
Digital Signal Processing
بسم الله الرحمن الرحيم Digital Signal Processing Lecture 2 Analog to Digital Conversion University of Khartoum Department of Electrical and Electronic.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate accuracy.
Appendix A: Probability Theory
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
CH 8. Image Compression 8.1 Fundamental 8.2 Image compression models
Why Compress? To reduce the volume of data to be transmitted (text, fax, images) To reduce the bandwidth required for transmission and to reduce storage.
Image Transforms for Robust Coding
Sampling and Quantization
EE513 Audio Signals and Systems
Lecture 7 Sampling and Sampling Distributions
Chapter 8 Estimation.
Discrete Random Variables and Probability Distributions
Presentation transcript:

©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms

©2003/04 Alessandro Bogliolo Outline 1.Information theory 1.Definitions 2.Encodings 3.Digitalization 2.Probability theory 1.Random process 2.Probability (marginal, join, conditioned), Bayes Theorem 3.Probability distribution 4.Sampling theory and estimation 3.Algorithms 1.Definition 2.Equivalence 3.Complexity

©2003/04 Alessandro Bogliolo Information Information: reduction of uncertainty The minimum uncertainty is given by two alternatives The elementary choice between 2 alternatives contains the minimum amount of information Bit: binary digit encoding the elementary choice between 2 alternatives (information unit)

©2003/04 Alessandro Bogliolo Strings String: sequence of N characters (or digits) taken from a given finite alphabete of S symbols There are S N different strings of N characters taken from the same alphabete of S symbols There are S N different configurations of N characters taken from an alphabete of S symbols A binary string is composed of bits, defined over the binary alphabete Byte: binary string of 8 bits

©2003/04 Alessandro Bogliolo Encodings Encoding: assignment of strings with the elements of a set according to a given rule Properties: –Irredundant: each element is assigned with a unique string –Constant length: all code words are of the same length –Exact: all elements are encoded and there are no elements associated with the same string

©2003/04 Alessandro Bogliolo Encoding finite sets The minimum number of digits of a constant-length exact encoding of a set of M elements is so that Properties:

©2003/04 Alessandro Bogliolo Encoding unlimited sets (limitation) An unlimited set contains infinite elements Example: integer numbers Infinite sets cannot be exactly encoded In order to be digitally encoded the set must be restricted to a limited, finite subset In most cases this is done by encoding only the elements within given lower and upper bounds Example: integer numbers within 0 and 999 The limited subset may be exactly encoded

©2003/04 Alessandro Bogliolo Example: positional notation The base-b positional representation of an integer number of n digits has the form The value of the number is n digits encode all integer numbers from 0 to b n -1 Example: b=2, n=510011=1*16+1*2+1*1= =31=2 5 -1

©2003/04 Alessandro Bogliolo Encoding continuous sets (discretization) A continuous set contains infinite elements Example: real numbers in a given interval, points on a plane In order to be digitally encoded, the set needs to be discretized: partitioned into a discrete number of subsets Codewords will be associated with subsets The resulting encoding is approximated

©2003/04 Alessandro Bogliolo Example: gray levels The encoding associates a unique code with an interval of gray levels All gray levels within the interval are associated with the same code, thus loosing information The original gray level cannot be exactly reconstructed from the code Encoding associates each code with a unique gray level (representative of a class)

©2003/04 Alessandro Bogliolo Example: 2D images Gray level x y n lev nxnx nyny pixel

©2003/04 Alessandro Bogliolo Analog and digital signals Signal: time-varying physical quantity –Analog: continuous-time, continuous-value –Digital: discrete-time, discrete-value The digital encoding of a continuous signal entails: –Sampling (i.e., time discretization) –Quantization (i.e., value discretization) Sampling rate Duration Sample size

©2003/04 Alessandro Bogliolo Example: time series time value

©2003/04 Alessandro Bogliolo Example: video s rate = frame rate n col = number of colors n x n y = frame size time nyny nxnx color

©2003/04 Alessandro Bogliolo Redundancy Redundant encoding: encoding that makes use of more than the minimum number of digits required by an exact encoding Motivations for redundancy: –Providing more expressive/natural encoding/decoding rules –Reliability (error detection) Ex: parity encoding –Noise immunity / fault tolerance (error correction) Ex: triplication

©2003/04 Alessandro Bogliolo Parity encoding: –A parity bit is used to guarantee that all codewords have an even number of 1’s –Single errors are detected by means of a parity check Redundancy: examples parity check 0 1 error Irredundant codeword Triple redundancy: –Each character is repeats 3 times –Single errors are corrected by means of a majority voting error voting result

©2003/04 Alessandro Bogliolo Compression Lossy compression –Compression achieved at the cost of reducing the accuracy of the representation –The original representation cannot be restored Lossless compression –Compression achieved by either removing redundancy or leveraging content-specific opportunities –The original representation can be restored

©2003/04 Alessandro Bogliolo Outline 1.Information theory 1.Definitions 2.Encodings 3.Digitalization 2.Probability theory 1.Random process 2.Probability (marginal, join, conditioned), Bayes Theorem 3.Probability distribution 4.Sampling theory and estimation 3.Algorithms 1.Definition 2.Equivalence 3.Complexity

©2003/04 Alessandro Bogliolo Random process A random process: –Can be repeated infinite times –May provide mutually-exclusive results –Provide an unpredictable result at each trial Elementary event (e): each possible result of a single trial Event space (X): the set of all elementary events Event (E): any set of elementary events (any subset of the event space) E X e

©2003/04 Alessandro Bogliolo Probability Relative frequency of E over N trials: ratio of the number of occurrence of E (n E ) over the N trials: f N (E) = n E /N Probability of E (empirical definition): Probability of E (assiomatic definition): p(E)=0if E is empty p(E)=1if E=X if

©2003/04 Alessandro Bogliolo Probability (properties) 1.If all elementary events have the same probability, the probability of an event is given by its “relative size”: p(E) = card(E)/card(X) 2. 3.If E 1 is a subset of E E1E1 E2E2 X

©2003/04 Alessandro Bogliolo Conditional probability Joint probability of two events, outcomes of two random experiments Marginal probability Conditional probability Decomposition

©2003/04 Alessandro Bogliolo Independent events The joint probability of two independent events is equal to the product of their marginal probabilities E 1 and E 2 are independent events if and only if

©2003/04 Alessandro Bogliolo Bayes Theorem Bayes theorem: given two events E 1 and E 2, the conditional probability of E 1 given E 2 can be expressed as: The theorem provides the statistical support for statistical diagnosis based on the evaluation of the probability of a possible cause (E 1 ) of an observed effect (E 2 )

©2003/04 Alessandro Bogliolo Random variable Random variable x: variable representing the outcome of a random experiment Probability distribution function of x: Probability density function of x:

©2003/04 Alessandro Bogliolo Sampling and estimation Parent population of a random variable x: ideal set of the outcomes of all possible trials of a random process Sample of x: set of the outcomes of N trials of the random process Sample parameters can be used as estimators of parameters of the parent population Example: Expected value of xSample average

©2003/04 Alessandro Bogliolo Confidence of an estimator The quality of the estimator P’ of a given parameter P can be expressed in terms of: –Confidence interval: limiting distance d between the estimator and the actual parameter –Confidence level: probability c of finding the actual parameter within the confidence interval The smaller the confidence interval d and the higher the confidence level c, the better The quality of an estimator grows with the number of samples For a fixed confidence level c, the size of confidence interval d decreases with the inverse of the square root of N

©2003/04 Alessandro Bogliolo Variance, covariance, correlation Variance: Standard deviation: Covariance: Correlation: The confidence interval of an estimator is proportional to

©2003/04 Alessandro Bogliolo Outline 1.Information theory 1.Definitions 2.Encodings 3.Digitalization 2.Probability theory 1.Random process 2.Probability (marginal, join, conditioned), Bayes Theorem 3.Probability distribution 4.Sampling theory and estimation 3.Algorithms 1.Definition 2.Equivalence 3.Complexity

©2003/04 Alessandro Bogliolo Algorithm Definition: Finite description of a finite sequence of non ambiguous instructions that can be executed in finite time to solve a problem or provide a result Key properties: –Finite description –Non ambiguity –Finite execution Algorithms take input data and provide output data Domain: set of all allowed configurations of the input data

©2003/04 Alessandro Bogliolo Complexity Complexity: measure of the number of elementary steps required by the algorithm to solve a problem The number of execution steps usually depend on the configuration of the input data (i.e., on the instance of the problem) The complexity of an algorithm is usually expressed as a function of its input data, retaining the type of behavior while neglecting additive and multiplicative constants Example: O(n), O(n 2 ), O(2 n )

©2003/04 Alessandro Bogliolo Equivalence Two algorithms are said to be equivalent if: –they are defined on the same domain –they provide the same result in all domain points In general, there are many equivalent algorithms that solve the same problem, possibly providing different complexity The complexity is a property of an algorithm, it is not an inherent property of the problem The complexity of the most efficient knwon algorithm that solves a given problem is commonly considered to be the complexity of the problem