Random Variables & Entropy: Extension and Examples Brooks Zurn EE 270 / STAT 270 FALL 2007.

Slides:



Advertisements
Similar presentations
ELEN 5346/4304 DSP and Filter Design Fall Lecture 15: Stochastic processes Instructor: Dr. Gleb V. Tcheslavski Contact:
Advertisements

Probabilistic models Haixu Tang School of Informatics.
Random Variables ECE460 Spring, 2012.
Acknowledgement: Thanks to Professor Pagano
Important Random Variables Binomial: S X = { 0,1,2,..., n} Geometric: S X = { 0,1,2,... } Poisson: S X = { 0,1,2,... }
Random Variable A random variable X is a function that assign a real number, X(ζ), to each outcome ζ in the sample space of a random experiment. Domain.
Today Today: Finish Chapter 4, Start Chapter 5 Reading: –Chapter 5 (not 5.12) –Important Sections From Chapter (excluding the negative hypergeometric.
Review of Basic Probability and Statistics
Introduction to stochastic process
Chapter 1 Probability Theory (i) : One Random Variable
Review.
1 Chap 5 Sums of Random Variables and Long-Term Averages Many problems involve the counting of number of occurrences of events, computation of arithmetic.
Prof. Bart Selman Module Probability --- Part d)
Probability Distributions
1 Review of Probability Theory [Source: Stanford University]
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Probability and Statistics Review
A random variable that has the following pmf is said to be a binomial random variable with parameters n, p The Binomial random variable.
The moment generating function of random variable X is given by Moment generating function.
Ya Bao Fundamentals of Communications theory1 Random signals and Processes ref: F. G. Stremler, Introduction to Communication Systems 3/e Probability All.
Class notes for ISE 201 San Jose State University
Discrete Random Variables and Probability Distributions
Discrete and Continuous Distributions G. V. Narayanan.
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
Hamid R. Rabiee Fall 2009 Stochastic Processes Review of Elementary Probability Lecture I.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
Short Resume of Statistical Terms Fall 2013 By Yaohang Li, Ph.D.
Chapter 5 Statistical Models in Simulation
Tch-prob1 Chap 3. Random Variables The outcome of a random experiment need not be a number. However, we are usually interested in some measurement or numeric.
IRDM WS Chapter 2: Basics from Probability Theory and Statistics 2.1 Probability Theory Events, Probabilities, Random Variables, Distributions,
Winter 2006EE384x1 Review of Probability Theory Review Session 1 EE384X.
Dept of Bioenvironmental Systems Engineering National Taiwan University Lab for Remote Sensing Hydrology and Spatial Modeling STATISTICS Random Variables.
TELECOMMUNICATIONS Dr. Hugh Blanton ENTC 4307/ENTC 5307.
2.1 Random Variable Concept Given an experiment defined by a sample space S with elements s, we assign a real number to every s according to some rule.
Further distributions
Statistics for Engineer Week II and Week III: Random Variables and Probability Distribution.
Random Variables and Stochastic Processes –
The Negative Binomial Distribution An experiment is called a negative binomial experiment if it satisfies the following conditions: 1.The experiment of.
Theory of Probability Statistics for Business and Economics.
Random Variables. A random variable X is a real valued function defined on the sample space, X : S  R. The set { s  S : X ( s )  [ a, b ] is an event}.
ENGR 610 Applied Statistics Fall Week 3 Marshall University CITE Jack Smith.
Math 4030 – 4a Discrete Distributions
King Saud University Women Students
Probability Review-1 Probability Review. Probability Review-2 Probability Theory Mathematical description of relationships or occurrences that cannot.
Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back.
IE 300, Fall 2012 Richard Sowers IESE. 8/30/2012 Goals: Rules of Probability Counting Equally likely Some examples.
Review of Chapter
4.3 More Discrete Probability Distributions NOTES Coach Bridges.
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
Discrete Random Variables. Introduction In previous lectures we established a foundation of the probability theory; we applied the probability theory.
Chapter 6: Continuous Probability Distributions A visual comparison.
Chapter 31Introduction to Statistical Quality Control, 7th Edition by Douglas C. Montgomery. Copyright (c) 2012 John Wiley & Sons, Inc.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Engineering Probability and Statistics - SE-205 -Chap 3 By S. O. Duffuaa.
Chapter 6: Continuous Probability Distributions A visual comparison.
Analytical Model for Connectivity in Vehicular Ad Hoc Networks Published in IEEE TVT vol. 57, no. 6, November 2008 Authors: Saleh Yousefi, Eitan Altman,
Probability & Random Variables
Basic statistics Usman Roshan.
Review of Probability Theory
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Statistical Modelling
Random variables (r.v.) Random variable
Basic statistics Usman Roshan.
Probability for Machine Learning
Appendix A: Probability Theory
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Probability Review for Financial Engineers
Chapter 3 : Random Variables
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Presentation transcript:

Random Variables & Entropy: Extension and Examples Brooks Zurn EE 270 / STAT 270 FALL 2007

Overview Density Functions and Random Variables Distribution Types Entropy

Density Functions PDF vs. CDF – PDF shows probability of each size bin – CDF shows cumulative probability for all sizes up to and including current bin – This data shows the normalized, relative size of a rodent as seen from an overhead camera for 8 behaviors

Markov & Chebyshev Inequalities What’s the point? Setting a maximum limit on probability This limits the search space for a solution – When looking for a needle in a haystack, it helps to have a smaller haystack. Can use limit to determine the necessary sample size

Markov & Chebyshev Inequalities Example: Mean height of a child in a kindergarten class is 3’6”. (Leon-Garcia text, p. 137 – see end of presentation) – Using Markov’s inequality, the probability of a child being taller than 9 feet is <= 42/108 =.389.  there will be fewer than 39 students over 9 feet tall in a class of 100 students.  Also, there will be NO LESS THAN 41 students who are under 9’ tall. -Using Chebyshev’s inequality (and assuming the variance = 1 foot) the probability of a child being taller than 9 feet is <= 12 2 /108 2 =  there will be no more than 2 students taller than 9’ in a class of 100 students. (this is also consistent with Markov’s Inequality).  Also, there will be NO LESS THAN 98 students under 9’ tall. This gives us a basic idea of how many student heights we need to measure to rule out the possibility that we have a 9’ tall student… SAMPLE SIZE!!

Markov’s Inequality Derivation: For a random variable X >= 0, E[x]=, where f x (x)=P[x-e/2£X£x+e/2]/e Assuming this also holds for X = a, because this is a continuous integral.

Markov’s Inequality Therefore for c > 0, the number of values of x > c is infinite, therefore the value of c will stay constant while x continues to increase.

Markov’s Inequality References: Lefebvre text.

Chebyshev’s Inequality Derivation (INCOMPLETE):

Chebyshev’s Inequality As before, c 2 is constant and (Y-E[Y]) 2 continues to increase. But, how do f y |Y-E[Y]| and f Y (Y-E[Y]) 2 relate? (|Y-E[Y]|) 2 = (Y-E[Y]) 2 As long as Y – E[Y] is >= 1, then u 2 will be > u and the inequality holds, as per Markov’s Inequality. Note: this is not a rigorous proof, and cases for which Y – E[Y] < 1 are not discussed. Reference: Lefebvre text.

Note These both involve the Central Limit Theorem, which is derived in the Leon-Garcia text on p Central Limit Theorem states that the CDF of a normalized sequence of n random variables approaches the CDF of a Gaussian random variable. (p. 280)

Overview Entropy – What is it? – Used in…

Entropy What is it? – According to Jorge Cham (PhD Comics),

Entropy “Measure of uncertainty in a random experiment” Reference: Leon-Garcia Text Used in information theory – Message transmission (for example, Lathi text p. 682) – Decision Tree ‘Gain Criterion’ Leon-Garcia text p. 167 ID3, C4.5, ITI, etc. by J. Ross Quinlan and Paul Utgoff Note: NOT same as the Gini index used as a splitting criterion by the CART tree method (Breiman et al, 1984).

Entropy ID3 Decision Tree: Expected Information for a Binary Tree where the entropy I is E(A) is the average information needed to classify A. ITI (Incremental Tree Inducer): - Based on ID3 and its successor, C4.5. -Uses a gain ratio metric to improve function for certain cases

Entropy ITI Decision Tree for Rodent Behaviors – ITI is an extension of ID3 Reference: ‘Rodent Data’ paper.

Distribution Types Continuous Random Variables – Normal (or Gaussian) Distribution – Uniform Distribution – Exponential Distribution – Rayleigh Random Variable Discrete (‘counting’) Random Variables – Binomial Distribution – Bernoulli and Geometric Distributions – Poisson Distribution

Poisson Distribution Number of events occurring in one time unit, time between events is exponentially distributed with mean 1/a. Gives a method for modeling completely random, independent events that occur after a random interval of time. (Leon-Garcia p. 106) Poisson Dist. can model a sequence of Bernoulli trials (Leon-Garcia p. 109) – Bernoulli gives the probability of a single coin toss. and References: Kao text, Leon-Garcia text.

Poisson Distribution

References Lefebvre Text: – Applied Stochastic Processes, Mario Lefebvre. New York, NY: Springer., 2003 Kao Text: – An Introduction to Stochastic Processes, Edward P. C. Kao. Belmont, CA, USA: Duxbury Press at Wadsworth Publishing Company, Lathi Text: – Modern Digital and Analog Communication Systems, 3 rd ed., B. P. Lathi. New York, Oxford: Oxford University Press, Entropy-Based Decision Trees: – ID3: P. E. Utgoff, "Incremental induction of decision trees.," Machine Learning, vol. 4, pp , – C4.5: J. R. Quinlan, C4.5: Programs for machine learning, 1st ed. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc., – ITI: P. E. Utgoff, N. C. Berkman, and J. A. Clouse, "Decision tree induction based on efficient tree restructuring.," Machine Learning, vol. 29, pp. 5-44, Other Decision Tree Methods: – CART: L. Breiman, J. H. Friedman, R. A. Olshen, C. J. Stone, Classification and Regression Trees. Belmont, CA: Wadsworth Rodent Data: – J. Brooks Zurn, Xianhua Jiang, Yuichi Motai. Video-Based Tracking and Incremental Learning Applied to Rodent Behavioral Activity under Near-Infrared Illumination. To appear: IEEE Transactions on Instrumentation and Measurement, December 2007 or February Poisson Distribution Example: –

Questions?