Presented by Minkoo Seo March, 2006

Slides:



Advertisements
Similar presentations
Entropy and Information Theory
Advertisements

Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
CS433: Modeling and Simulation
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
SIMS-201 Compressing Information. 2  Overview Chapter 7: Compression Introduction Entropy Huffman coding Universal coding.
Chain Rules for Entropy
Measures of Information Hartley defined the first information measure: –H = n log s –n is the length of the message and s is the number of possible values.
QBM117 Business Statistics
Review of Basic Probability and Statistics
Mutual Information Mathematical Biology Seminar
A Bit of Information Theory Unsupervised Learning Working Group Assaf Oron, Oct Based mostly upon: Cover & Thomas, “Elements of Inf. Theory”,
Fundamental limits in Information Theory Chapter 10 :
Ref. Cryptography: theory and practice Douglas R. Stinson
Information Theory and Security. Lecture Motivation Up to this point we have seen: –Classical Crypto –Symmetric Crypto –Asymmetric Crypto These systems.
1 Chapter 5 A Measure of Information. 2 Outline 5.1 Axioms for the uncertainty measure 5.2 Two Interpretations of the uncertainty function 5.3 Properties.
Information Theory Rong Jin. Outline  Information  Entropy  Mutual information  Noisy channel model.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
Information Theory and Security
Probability and Statistics Review Thursday Sep 11.
Bioinformatics lectures at Rice University Lecture 4: Shannon entropy and mutual information.
Chapter 16: Random Variables
1 Statistical NLP: Lecture 5 Mathematical Foundations II: Information Theory.
Some basic concepts of Information Theory and Entropy
Information Theory and Security Prakash Panangaden McGill University First Canada-France Workshop on Foundations and Practice of Security Montréal 2008.
Information and Coding Theory
§1 Entropy and mutual information
STATISTIC & INFORMATION THEORY (CSNB134)
2. Mathematical Foundations
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Chapter 4 Probability See.
1 CY1B2 Statistics Aims: To introduce basic statistics. Outcomes: To understand some fundamental concepts in statistics, and be able to apply some probability.
Discrete Random Variables: PMFs and Moments Lemon Chapter 2 Intro to Probability
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
Chapter 16: Random Variables
Prof. Pushpak Bhattacharyya, IIT Bombay1 Basics Of Entropy CS 621 Artificial Intelligence Lecture /09/05 Prof. Pushpak Bhattacharyya.
Summer 2004CS 4953 The Hidden Art of Steganography A Brief Introduction to Information Theory  Information theory is a branch of science that deals with.
2.1 Introduction In an experiment of chance, outcomes occur randomly. We often summarize the outcome from a random experiment by a simple number. Definition.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 27, 2012.
Coding Theory Efficient and Reliable Transfer of Information
CS433 Modeling and Simulation Lecture 03 – Part 01 Probability Review 1 Dr. Anis Koubâa Al-Imam Mohammad Ibn Saud University
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Measuring chance Probabilities FETP India. Competency to be gained from this lecture Apply probabilities to field epidemiology.
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
AP STATISTICS Section 7.1 Random Variables. Objective: To be able to recognize discrete and continuous random variables and calculate probabilities using.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Decision Trees.
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
Probability Michael J. Watts
Lecture 3 Appendix 1 Computation of the conditional entropy.
Oliver Schulte Machine Learning 726 Decision Tree Classifiers.
Introduction A probability distribution is obtained when probability values are assigned to all possible numerical values of a random variable. It may.
Mutual Information and Channel Capacity Multimedia Security.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Chapter 4: Information Theory. Learning Objectives LO 4.1 – Understand discrete and continuous messages, message sources, amount of information and its.
Information Theory Information Suppose that we have the source alphabet of q symbols s 1, s 2,.., s q, each with its probability p(s i )=p i. How much.
Random Variables and Probability Distributions. Definition A random variable is a real-valued function whose domain is the sample space for some experiment.
Chapter 6 – Continuous Probability Distribution Introduction A probability distribution is obtained when probability values are assigned to all possible.
Statistical methods in NLP Course 2 Diana Trandab ă ț
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Statistical methods in NLP Course 2
Shannon Entropy Shannon worked at Bell Labs (part of AT&T)
Information Theory Michael J. Watts
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
Means and Variances of Random Variables
Information Based Criteria for Design of Experiments
A Brief Introduction to Information Theory
Problem 1.
Entropy CSCI284/162 Spring 2009 GWU.
Chapter 5 – Probability Rules
Presentation transcript:

Presented by Minkoo Seo March, 2006 Entropy Presented by Minkoo Seo March, 2006

Self information (Surprisal) Suppose that almost all children like ice cream. You ask her, “Do you like ice cream?” Her answer is always yes. There’s no surprisal and no information.

Self information (Surprisal) (cont) If a child answers, “No”, then you are surprised. You expect more information from surprising answers. In other words,

Self information (Surprisal) (cont) Now, suppose that there are two events. We want to get ‘self information’ value from them. The result would be

Self information (Surprisal) (cont) Multiplication is neither easy nor intuitive. We want information can be added like other measures. So, information Probability of a message

Entropy Suppose that you are tossing a coin. Your observation was TTHTHTTT. Then, how can we measure the average information per symbol?

Entropy (cont) We define Shannon’s entropy, or decrease in uncertainty, as Expected self information All messages

Joint Entropy Joint entropy of two discrete random variable is defined as

Conditional Entropy Conditional entropy of a discrete random variable X given Y=y is defined as Basic property of the conditional entropy Information of X when we know Y Information of X,Y Given Y

Mutual Information A quantity that measures the mutual dependence of the two variables. How much information can be obtained by observing the other.

Mutual Information (cont) Definition Amount of uncertainty of B(or A) reduced when A(or B) is known Information of B gained by observing A Information of B Information of B which is remained when we know A

Applications Coding theory Clustering Bits needed for describing a data Clustering Make a cluster from two variables whose mutual information is large Large mutual information implies that those two variables conveys the almost same information

References Wikipedia Ian Korf et al, BLAST, O’Reilly, 2003.