Independence of random variables

Slides:



Advertisements
Similar presentations
Random Processes Introduction (2)
Advertisements

Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Section 7.4 (partially). Section Summary Expected Value Linearity of Expectations Independent Random Variables.
April 2, 2015Applied Discrete Mathematics Week 8: Advanced Counting 1 Random Variables In some experiments, we would like to assign a numerical value to.
Chapter 2: Probability Random Variable (r.v.) is a variable whose value is unknown until it is observed. The value of a random variable results from an.
Chapter 4 Discrete Random Variables and Probability Distributions
Class notes for ISE 201 San Jose State University
Prof. Bart Selman Module Probability --- Part d)
The moment generating function of random variable X is given by Moment generating function.
Chapter 4: Joint and Conditional Distributions
Week 51 Theorem For g: R  R If X is a discrete random variable then If X is a continuous random variable Proof: We proof it for the discrete case. Let.
Review of Probability and Statistics
The joint probability distribution function of X and Y is denoted by f XY (x,y). The marginal probability distribution function of X, f X (x) is obtained.
Random Variable and Probability Distribution
5-1 Two Discrete Random Variables Example Two Discrete Random Variables Figure 5-1 Joint probability distribution of X and Y in Example 5-1.
Joint Probability distribution
5-1 Two Discrete Random Variables Example Two Discrete Random Variables Figure 5-1 Joint probability distribution of X and Y in Example 5-1.
NIPRL Chapter 2. Random Variables 2.1 Discrete Random Variables 2.2 Continuous Random Variables 2.3 The Expectation of a Random Variable 2.4 The Variance.
Joint Distribution of two or More Random Variables
Definition of Covariance The covariance of X & Y, denoted Cov(X,Y), is the number where  X = E(X) and  Y = E(Y). Computational Formula:
Section 8 – Joint, Marginal, and Conditional Distributions.
1 Lecture 4. 2 Random Variables (Discrete) Real-valued functions defined on a sample space are random vars. determined by outcome of experiment, we can.
Lectures prepared by: Elchanan Mossel elena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
0 K. Salah 2. Review of Probability and Statistics Refs: Law & Kelton, Chapter 4.
1 G Lect 4a G Lecture 4a f(X) of special interest: Normal Distribution Are These Random Variables Normally Distributed? Probability Statements.
Convergence in Distribution
Discrete Random Variables A random variable is a function that assigns a numerical value to each simple event in a sample space. Range – the set of real.
Chapter 4 DeGroot & Schervish. Variance Although the mean of a distribution is a useful summary, it does not convey very much information about the distribution.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
STA347 - week 51 More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function.
Week 21 Conditional Probability Idea – have performed a chance experiment but don’t know the outcome (ω), but have some partial information (event A) about.
Expectation for multivariate distributions. Definition Let X 1, X 2, …, X n denote n jointly distributed random variable with joint density function f(x.
1 G Lect 2M Examples of Correlation Random variables and manipulated variables Thinking about joint distributions Thinking about marginal distributions:
STA347 - week 31 Random Variables Example: We roll a fair die 6 times. Suppose we are interested in the number of 5’s in the 6 rolls. Let X = number of.
Statistics for Business & Economics
Consistency An estimator is a consistent estimator of θ, if , i.e., if
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
Week 121 Law of Large Numbers Toss a coin n times. Suppose X i ’s are Bernoulli random variables with p = ½ and E(X i ) = ½. The proportion of heads is.
Lecture 3 1 Recap Random variables Continuous random variable Sample space has infinitely many elements The density function f(x) is a continuous function.
Chapter Eight Expectation of Discrete Random Variable
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
1 Probability and Statistical Inference (9th Edition) Chapter 4 Bivariate Distributions November 4, 2015.
Section 10.5 Let X be any random variable with (finite) mean  and (finite) variance  2. We shall assume X is a continuous type random variable with p.d.f.
Week 111 Some facts about Power Series Consider the power series with non-negative coefficients a k. If converges for any positive value of t, say for.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Chebyshev’s Inequality Markov’s Inequality Proposition 2.1.
Chapter 5 Joint Probability Distributions and Random Samples  Jointly Distributed Random Variables.2 - Expected Values, Covariance, and Correlation.3.
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
1 Chapter 4 Mathematical Expectation  4.1 Mean of Random Variables  4.2 Variance and Covariance  4.3 Means and Variances of Linear Combinations of Random.
Conditional Expectation
Function of a random variable Let X be a random variable in a probabilistic space with a probability distribution F(x) Sometimes we may be interested in.
Week 61 Poisson Processes Model for times of occurrences (“arrivals”) of rare phenomena where λ – average number of arrivals per time period. X – number.
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Chapter 3: Discrete Random Variables and Their Distributions CIS.
Inequalities, Covariance, examples
Lebesgue measure: Lebesgue measure m0 is a measure on i.e., 1. 2.
PRODUCT MOMENTS OF BIVARIATE RANDOM VARIABLES
Parameter, Statistic and Random Samples
Conditional Probability on a joint discrete distribution
Chapter 4: Mathematical Expectation:
Monte Carlo Approximations – Introduction
Multinomial Distribution
How accurately can you (1) predict Y from X, and (2) predict X from Y?
Independence of random variables
Handout Ch 4 實習.
Handout Ch 4 實習.
9. Limit Theorems.
Mathematical Expectation
Presentation transcript:

Independence of random variables Definition Random variables X and Y are independent if their joint distribution function factors into the product of their marginal distribution functions Theorem Suppose X and Y are jointly continuous random variables. X and Y are independent if and only if given any two densities for X and Y their product is the joint density for the pair (X,Y) i.e. Proof: If X and Y are independent random variables and Z =g(X), W = h(Y) then Z, W are also independent. week 8

Example Suppose X and Y are discrete random variables whose values are the non-negative integers and their joint probability function is Are X and Y independent? What are their marginal distributions? Factorization is enough for independence, but we need to be careful of constant terms for factors to be marginal probability functions. week 8

Example and Important Comment The joint density for X, Y is given by Are X, Y independent? Independence requires that the set of points where the joint density is positive must be the Cartesian product of the set of points where the marginal densities are positive i.e. the set of points where fX,Y(x,y) >0 must be (possibly infinite) rectangles. week 8

Conditional densities If X, Y jointly distributed continuous random variables, the conditional density function of Y | X is defined to be if fX(x) > 0 and 0 otherwise. If X, Y are independent then . Also, Integrating both sides over x we get This is a useful application of the law of total probability for the continuous case. week 8

Example Consider the joint density Find the conditional density of X given Y and the conditional density of Y given X. week 8

Properties of Expectations Involving Joint Distributions For random variables X, Y and constants E(aX + bY) = aE(X) + bE(Y) Proof: For independent random variables X, Y E(XY) = E(X)E(Y) whenever these expectations exist. week 8

Covariance Recall: Var(X+Y) = Var(X) + Var(Y) +2 E[(X-E(X))(Y-E(Y))] Definition For random variables X, Y with E(X), E(Y) < ∞, the covariance of X and Y is Covariance measures whether or not X-E(X) and Y-E(Y) have the same sign. Claim: Proof: Note: If X, Y independent then E(XY) =E(X)E(Y), and Cov(X,Y) = 0. week 8

Example Suppose X, Y are discrete random variables with probability function given by Find Cov(X,Y). Are X,Y independent? y x -1 1 pX(x) 1/8 pY(y) week 8

Important Facts Independence of X, Y implies Cov(X,Y) = 0 but NOT vice versa. If X, Y independent then Var(X+Y) = Var(X) + Var(Y). If X, Y are NOT independent then Var(X+Y) = Var(X) + Var(Y) + 2Cov(X,Y). Cov(X,X) = Var(X). week 8

Example Suppose Y ~ Binomial(n, p). Find Var(Y). week 8

Properties of Covariance For random variables X, Y, Z and constants Cov(aX+b, cY+d) = acCov(X,Y) Cov(X+Y, Z) = Cov(X,Z) + Cov(Y,Z) Cov(X,Y) = Cov(Y, X) week 8

Correlation Definition For X, Y random variables the correlation of X and Y is whenever V(X), V(Y) ≠ 0 and all these quantities exists. Claim: ρ(aX+b,cY+d) = ρ(X,Y) Proof: This claim means that the correlation is scale invariant. week 8

Theorem For X, Y random variables, whenever the correlation ρ(X,Y) exists it must satisfy -1 ≤ ρ(X,Y) ≤ 1 Proof: week 8

Interpretation of Correlation ρ ρ(X,Y) is a measure of the strength and direction of the linear relationship between X, Y. If X, Y have non-zero variance, then . If X, Y independent, then ρ(X,Y) = 0. Note, it is not the only time when ρ(X,Y) = 0 !!! Y is a linearly increasing function of X if and only if ρ(X,Y) = 1. Y is a linearly decreasing function of X if and only if ρ(X,Y) = -1. week 8

Example Find Var(X - Y) and ρ(X,Y) if X, Y have the following joint density week 8

Conditional Expectation For X, Y discrete random variables, the conditional expectation of Y given X = x is and the conditional variance of Y given X = x is where these are defined only if the sums converges absolutely. In general, week 8

For X, Y continuous random variables, the conditional expectation of Y given X = x is and the conditional variance of Y given X = x is In general, week 8

Example Suppose X, Y are continuous random variables with joint density function Find E(X | Y = 2). week 8

More about Conditional Expectation Assume that E(Y | X = x) exists for every x in the range of X. Then, E(Y | X ) is a random variable. The expectation of this random variable is E [E(Y | X )] Theorem E [E(Y | X )] = E(Y) This is called the “Law of Total Expectation”. Proof: week 8

Example Suppose we roll a fair die; whatever number comes up we toss a coin that many times. What is the expected number of heads? week 8

Theorem For random variables X, Y V(Y) = V [E(Y|X)] + E[V(Y|X)] Proof: week 8

Example Let X ~ Geometric(p). Given X = x, let Y have conditionally the Binomial(x, p) distribution. Scenario: doing Bernoulli trails with success probability p until 1st success so X : number of trails. Then do x more trails and count the number of success which is Y. Find, E(Y), V(Y). week 8

Law of Large Numbers Toss a coin n times. Suppose Xi’s are Bernoulli random variables with p = ½ and E(Xi) = ½. The proportion of heads is . Intuitively approaches ½ as n  ∞ . week 8

Markov’s Inequality If X is a non-negative random variable with E(X) < ∞ and a >0 then, Proof: week 8

Chebyshev’s Inequality For a random variable X with E(X) < ∞ and V(X) < ∞, for any a >0 Proof: week 8

Back to the Law of Large Numbers Interested in sequence of random variables X1, X2, X3,… such that the random variables are independent and identically distributed (i.i.d). Let Suppose E(Xi) = μ , V(Xi) = σ2, then and Intuitively, as n  ∞, so week 8

Formally, the Weak Law of Large Numbers (WLLN) states the following: Suppose X1, X2, X3,…are i.i.d with E(Xi) = μ < ∞ , V(Xi) = σ2 < ∞, then for any positive number a as n  ∞ . This is called Convergence in Probability. Proof: week 8

Example Flip a coin 10,000 times. Let E(Xi) = ½ and V(Xi) = ¼ . Take a = 0.01, then by Chebyshev’s Inequality Chebyshev Inequality gives a very weak upper bound. Chebyshev Inequality works regardless of the distribution of the Xi’s. week 8

Strong Law of Large Number Suppose X1, X2, X3,…are i.i.d with E(Xi) = μ < ∞ , then converges to μ as n  ∞ with probability 1. That is This is called convergence almost surely. week 8