Inequalities, Covariance, examples

Slides:



Advertisements
Similar presentations
Joint Probability Distributions and Random Samples
Advertisements

Joint and marginal distribution functions For any two random variables X and Y defined on the same sample space, the joint c.d.f. is For an example, see.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Random Variable A random variable X is a function that assign a real number, X(ζ), to each outcome ζ in the sample space of a random experiment. Domain.
Independence of random variables
Class notes for ISE 201 San Jose State University
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Chapter 6 Continuous Random Variables and Probability Distributions
Chapter 4: Joint and Conditional Distributions
Chapter 5 Continuous Random Variables and Probability Distributions
Review of Probability and Statistics
Random Variable and Probability Distribution
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Functions of Two Random.
1 G Lect 4a G Lecture 4a f(X) of special interest: Normal Distribution Are These Random Variables Normally Distributed? Probability Statements.
MA-250 Probability and Statistics Nazar Khan PUCIT Lecture 26.
Convergence in Distribution
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Expectation. Let X denote a discrete random variable with probability function p(x) (probability density function f(x) if X is continuous) then the expected.
Week 121 Law of Large Numbers Toss a coin n times. Suppose X i ’s are Bernoulli random variables with p = ½ and E(X i ) = ½. The proportion of heads is.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
1 Probability and Statistical Inference (9th Edition) Chapter 4 Bivariate Distributions November 4, 2015.
Joint Moments and Joint Characteristic Functions.
Section 10.5 Let X be any random variable with (finite) mean  and (finite) variance  2. We shall assume X is a continuous type random variable with p.d.f.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.
More on Exponential Distribution, Hypo exponential distribution
Expectations of Random Variables, Functions of Random Variables
Virtual University of Pakistan
Chapter 5 Joint Probability Distributions and Random Samples
Some problems on Joint Distributions,
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Chapter 3: Discrete Random Variables and Their Distributions CIS.
ECE 313 Probability with Engineering Applications Lecture 7
Normal Distribution and Parameter Estimation
Expectations of Random Variables, Functions of Random Variables
Expectations of Random Variables, Functions of Random Variables
ECE 313 Probability with Engineering Applications Lecture 19
PRODUCT MOMENTS OF BIVARIATE RANDOM VARIABLES
Two Discrete Random Variables
Parameter, Statistic and Random Samples
Review of Probability and Estimators Arun Das, Jason Rebello
Some Rules for Expectation
Monte Carlo Approximations – Introduction
Sampling Distribution Models
Multinomial Distribution
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Random Variable X, with pmf p(x) or pdf f(x)
How accurately can you (1) predict Y from X, and (2) predict X from Y?
STOCHASTIC HYDROLOGY Random Processes
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
Virtual University of Pakistan
Independence of random variables
Handout Ch 4 實習.
Handout Ch 4 實習.
3. Random Variables Let (, F, P) be a probability model for an experiment, and X a function that maps every to a unique point.
9. Limit Theorems.
Chapter 5 Applied Statistics and Probability for Engineers
Chapter 2. Random Variables
9. Two Functions of Two Random Variables
Further Topics on Random Variables: Covariance and Correlation
Quick Review of Probability
Continuous Distributions
Further Topics on Random Variables: Covariance and Correlation
Applied Statistics and Probability for Engineers
Moments of Random Variables
Mathematical Expectation
Presentation transcript:

Inequalities, Covariance, examples ECE 313 Probability with Engineering Applications Lecture 25 Professor Ravi K. Iyer Dept. of Electrical and Computer Engineering University of Illinois at Urbana Champaign

Today’s Topics Final Exam on May 11, 8 am – 11 am. CLT, Inequalities Covariance Correlations: Announcements: Final Exam on May 11, 8 am – 11 am. Three 8x11 sheets allowed: No other aids electronic or otherwise Exam Review session May 9, 5:30pm Place TBD HW 10 deadline extended to Monday, May 1 Final project  MP3 Task 1 and Task 2 due today Wednesday, Apr 26 11;59pm  Final Presentation on Saturday, May 6, 12pm – 5pm Final Report will be due on Monday, May 8, 11:59pm

Final Project Presentation Format Presentation ~ 10 Slides Max: 2-3 Slides: Project Summary Patients and features selected and justify you choice. Why do you think your selected features would perform best on the complete patient set What concepts from class used for analysis, any additional concepts? Division of tasks 3-5 Slides: Analysis Results Provide evidence for the techniques you used. For example show the results of your analysis for selection of features in the form of tables/graphs presenting the ML and MAP errors, correlations, etc. Did you change your features selected in Task 2 after doing Task 3? Provide insights and justification for your selection. 1-2 Slides: Conclusions and key insights Example insights: Compare the results generated by ML and MAP rules. How were the projects useful in understanding the concepts learned in the class in practice? What suggestions do you have for improvement? DEMO ~ Groups should run their Matlab code for Task 3, and present the performance for the ML and MAP rule.

Limit Theorems: Strong Law of Large Numbers The strong law of large numbers: The average of a sequence of independent random variables having the same distribution will, with probability 1, converge to the mean of that distribution. Let X1,X2,… be a sequence of independent random variables having a common distribution, and let E[Xi]=μ Then, with probability 1, The central limit theorem provides a simple method for computing approximate probabilities for sums/means of independent random variables. It also explains the remarkable fact that the empirical frequencies of so many natural “populations” exhibit a bell-shaped (that is, normal) curve.

Limit Theorems: Central Limit Theorem Central Limit Theorem: Let X1, X2,…be a sequence of independent, identically distributed random variables, each with mean μ and variance σ2 then the distribution of Tends to the standard normal as . That is, Note that like other results, this theorem holds for any distribution of the Xi ‘s; herein lies its power.

Markov Inequality Proof: X is continuous with density f. Proposition - Markov’s Inequality: If X is a random variable that takes only nonnegative values, then for any value a>0 Proof: X is continuous with density f.

Chebychev’s Inequality As a corollary, we obtain the following Proposition - Chebyshev’s Inequality: If X is a random variable with mean μ and variance σ2 then for any value k>0, Proof: Since (X- μ)2 is a nonnegative random variable, we can apply Markov’s inequality (with a=k2) to obtain Since if and only if is equivalent to And the proof is complete

Inequalities Example The importance of Markov’s and Chebyshev’s inequalities is that they enable us to derive bounds on probabilities when only the mean, or both the mean and the variance, of the probability distribution are known. If the actual distribution were known, then the desired probabilities could be exactly computed, and we would not need to resort to bounds. Example: Suppose we know that the number of items produced in a factory during a week is a random variable with mean 500. What can be said about the probability that this week’s production will be at least 1000? If the variance of a week’s production is known to equal 100, then what can be said about the probability that this week’s production will be between 400 and 600?

Inequalities Example (Cont’d) Let X be the number of items that will be produced in a week. By Markov’s inequality, By Chebyshev’s inequality, Hence, And so the probability that this week’s production will be between 400 and 600 is at least 0.99.

Covariance The covariance of any two random variables, X and Y, denoted by Cov(X,Y), is defined by Covariance generalizes variance, in the sense that Var(X) = Cov(X, X). If either X or Y has mean zero, then E[XY] = Cov(X, Y ). If X and Y are independent then it follows that Cov(X,Y) = 0. But the reverse is false: Cov(X,Y) = 0 means X and Y are uncorrelated, but it doesn’t imply that X and Y are independent.

Correlation Coefficient The correlation between two random variable X and Y is measured using the correlation coefficient: (ρX,Y is well defined if Var(X) > 0 and Var(Y ) > 0) If Cov(X,Y) = 0, X and Y are called uncorrelated, which implies that E[XY] = E[X]E[Y]. If Cov(X,Y) > 0, X and Y are positively correlated, Y tends to increase as X does. If Cov(X,Y) < 0, X and Y are negatively correlated, Y tends to decrease as X increases.

Covariance Example Example: The joint density function of X,Y is given as follows: Verify that the preceding is a joint density function. Find Cov(X,Y). Are X and Y uncorrelated? Are X and Y independent? To show that f(x,y) is a joint density function we need to show it is nonnegative, which is immediate and that

Covariance Example To obtain E(Y), note that the density function of Y is Thus Y is an exponential random variable with parameter 1 Compute E[X] and E[XY] as follows: Now, the expected value of an exponential random variable with parameter 1/y, and thus is equal to y. Consequently,

Covariance Also Integration by parts gives Consequently, So X and Y are positively correlated because Cov(X,Y) > 0. Since Cov(X,Y) ≠ 0, X and Y are not independent.

Properties of Covariance For any random variable X, Y, Z, and constant c, we have: Cov (X,X) = Var(X), Cov (X,Y) = Cov(Y,X), Cov (cX,Y) = cCov(X,y), Cov (X,Y+Z) = Cov(X,Y) + Cov(X,Z). Whereas the first three properties are immediate, the final one is easily proven as follows: The last property generalizes to give the following result:

Properties of Covariance A useful expression for the variance of the sum of random variables can be obtained from the preceding equation: If are independent random variables, then the above equation reduces to:

Problem 1 Let X and Y be two continuous random variables with joint density function: where c is a constant. Part A: What is the value of c? Part B: Find the marginal pdfs of X and Y. Part C: Are X and Y independent? Why? Part D: Find P{X + Y < 3}. Show your work on choosing the limits of integrals by marking the region of integral.  

Problem 1 Solution We first define the region on which the joint distribution function f(x,y) is defined:  

Problem 1 Solution

Problem 1 Solution

Problem 2 The joint density of X and Y is given by: Find the region on which f(x, y) is defined Find C. Find the density function of X. Find the density function of Y. Find E[X]. Find E[Y]. Find P(X+Y<3), for X > 0.

Problem 1 - Solution

Problem 2 - Solution In this solution, we will make use of the identity Which follows because , is the density function of a gamma random variable with parameters n+1 and λ and must thus integrate to 1. hence C=1/4

Problem 2 – Solution, Cont’d Since the joint density is nonzero only when y>x and y>-x, we have, for x>0, for x<0 u = y - x

Problem 2 – Solution, Cont’d X > 0 X < 0

Problem 2 – Solution, Cont’d To find , we first draw the line X + Y = 3. For X > 0 :The region of integral is the triangle highlighted below, which is the intersection of regions and :

Problem 2 – Solution, Cont’d So for X > 0:

Problem 2 – Solution, Cont’d Note that for X < 0 : The region of integral would be the intersection of regions and . Calculate P(X+Y<3), for X < 0 as homework.

Problem 3 At even time instants, a robot moves either +1cm or –1cm in the x-direction according to the outcome of a coin flip (heads: +1cm, tails: –1cm); At odd time instants, a robot moves similarly according to another coin flip in the y-direction. Assuming that the robot begins at the origin, let X and Y be the coordinates of the location of the robot after 2n time instants. (Assume that the coins are not fair, and are flipped independently from each other. P(head) = p for both coins) a) What are the values X and Y can take? b) What is the marginal pmf of X and Y? c) What is the joint pmf P(X, Y)? What assumption did you have to make? Justify why the assumption is valid.

Problem 3 – Solution