Lectures prepared by: Elchanan Mossel Yelena Shvets

Slides:



Advertisements
Similar presentations
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Advertisements

Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Independence of random variables
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Sections 4.1, 4.2, 4.3 Important Definitions in the Text:
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Chapter 4: Joint and Conditional Distributions
Review of Probability and Statistics
The joint probability distribution function of X and Y is denoted by f XY (x,y). The marginal probability distribution function of X, f X (x) is obtained.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Definition of Covariance The covariance of X & Y, denoted Cov(X,Y), is the number where  X = E(X) and  Y = E(Y). Computational Formula:
ELEC 303, Koushanfar, Fall’09 ELEC 303 – Random Signals Lecture 11 – Derived distributions, covariance, correlation and convolution Dr. Farinaz Koushanfar.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel elena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
1 G Lect 4a G Lecture 4a f(X) of special interest: Normal Distribution Are These Random Variables Normally Distributed? Probability Statements.
Chapter 4 DeGroot & Schervish. Variance Although the mean of a distribution is a useful summary, it does not convey very much information about the distribution.
Chapter 5.6 From DeGroot & Schervish. Uniform Distribution.
1 G Lect 2M Examples of Correlation Random variables and manipulated variables Thinking about joint distributions Thinking about marginal distributions:
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lecture 3 1 Recap Random variables Continuous random variable Sample space has infinitely many elements The density function f(x) is a continuous function.
Lecture 29 Dr. MUMTAZ AHMED MTH 161: Introduction To Statistics.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
1 Two Discrete Random Variables The probability mass function (pmf) of a single discrete rv X specifies how much probability mass is placed on each possible.
Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Probability Refresher
Joint Probability Distributions and Random Samples
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Handout Ch 4 實習.
PRODUCT MOMENTS OF BIVARIATE RANDOM VARIABLES
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Some Rules for Expectation
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Chapter 3: Getting the Hang of Statistics
Chapter 10: Covariance and Correlation
Multinomial Distribution
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
How accurately can you (1) predict Y from X, and (2) predict X from Y?
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
Lectures prepared by: Elchanan Mossel Yelena Shvets
Independence of random variables
Chapter 3: Getting the Hang of Statistics
Handout Ch 4 實習.
Handout Ch 4 實習.
Lectures prepared by: Elchanan Mossel Yelena Shvets
Financial Econometrics Fin. 505
Chapter 2. Random Variables
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Further Topics on Random Variables: Covariance and Correlation
Lectures prepared by: Elchanan Mossel Yelena Shvets
Chapter 10: Covariance and Correlation
Further Topics on Random Variables: Covariance and Correlation
Chapter 10: Covariance and Correlation
Mathematical Expectation
Presentation transcript:

Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Lectures prepared by: Elchanan Mossel Yelena Shvets Follows Jim Pitman’s book: Probability Sections 6.4

Do taller people make more money? Question: How can this be measured? - Ave (height) wage at 19 Ave (wage) height at 16 National Longitudinal Survey of Youth 1997 (NLSY97)

Definition of Covariance Cov (X,Y)= E[(X-mX)(Y – mY)] Alternative Formula Cov (X,Y)= E(XY) – E(X)E(Y) Variance of a Sum Var (X+Y)= Var (X) + Var (Y)+2 Cov (X,Y) Claim: Covariance is Bilinear

What does the sign of covariance mean? Look at Y = aX + b. Then: Cov(X,Y) = Cov(X,aX + b) = aVar(X). Ave(X) Ave(X) y y Ave(Y) Ave(Y) a>0 a<0 x x If a > 0, above the average in X goes with above the ave in Y. If a < 0, above the average n X goes with below the ave in Y. Cov(X,Y) = 0 means that there is no linear trend which connects X and Y.

Meaning of the value of Covariance Back to the National Survey of Youth study : the actual covariance was 3028 where height is inches and the wages in dollars. Question: Suppose we measured all the heights in centimeters, instead. There are 2.54 cm/inch? Question: What will happen to the covariance? Solution: So let HI be height in inches and HC be the height in centimeters, with W – the wages. Cov(HC,W) = Cov(2.54 HI,W) = 2.54 Cov (HI,W). So the value depends on the units and is not very informative!

Covariance and Correlation Define the correlation coefficient: Using the linearity of Expectation we get: Notice that r(aX+b, cY+d) = r(X,Y). This new quantity is independent of the change in scale and it’s value is quite informative.

Covariance and Correlation Properties of correlation:

Covariance and Correlation Claim: The correlation is always between –1 and +1 r = 1 iff Y = aX + b.

Correlation and Independence X & Y are uncorrelated iff any of the following hold Cov(X,Y) = 0, Corr(X,Y) = 0 E(XY) = E(X) E(Y). In particular, if X and Y are independent they are uncorrelated. X2 Example: Let X» N(0,1) and Y = X2, then Cov(XY) =E(XY) – E(X)E(Y) = E(X3) = 0, since the density is symmetric. X

Roll a dye N times. Let X be #1’s, Y be #2’s. Question: What is the correlation between X and Y? Solution: To compute the correlation directly from the multinomial distribution would be difficult. Let’s use a trick: Var(X+Y) = Var(X) + Var(Y) + 2Cov(X,Y). Since X+Y is just the number of 1’s or 2’s, X+Y»Binom(p1+p2,N). Var(X+Y) = (p1+p2)(1 - p1+p2) N. And X»Binom(p1,N), Y»Binom(p2,N), so Var(X) =p1(1-p1)N; Var(Y) = p2(1-p2)N.

Correlations in the Multinomial Distribution Hence Cov(X,Y) = (Var(X+Y) – Var(X) – Var(Y))/2 Cov(X,Y) = N((p1+p2)(1 - p1-p2) - p1(1-p1) -p2(1-p2))/2 = -N p1 p2 In our case p1 = p2 = 1/6, so r = 1/5. The formula holds for a general multinomial distribution.

Variance of the Sum of N Variables Var(åi Xi) = åi Var(Xi) + 2 åj<i Cov(Xi Xj) Proof: Var(åi Xi) = E[åi Xi – E(åj Xi) ]2 [åi Xi – E(åj Xi) ]2 = [åi (Xi –mi) ]2 = åi (Xi –mi) 2 + 2 åj<i (Xi –mi) (Xj –mj). Now take expectations and we have the result.

Variance of the Sample Average Let the population be a list of N numbers x(1), …, x(N). Then are the population mean and population variance. Let X1, X2,…, Xn be a sample of size n drawn from this population. Then each Xk has the same distribution as the entire population and Let be the sample average.

Variance of the Sample Average By linearity of expectation , both for a sample drawn with and without replacement. When X1, X2,…, Xn are drawn with replacement, they are independent and each Xk has variance s2 . Then

Variance of the Sample Average Question: What is the SD for sampling without replacement? Solution: Let Sn = X1 + X2 + … + Xn. Then . Var(Sn) = åi Var(Xi) + 2 åj<i Cov(Xi Xj) By symmetry Cov(Xi,Xj) = Cov(X1,X2), so Var(Sn) = ns2 + n(n-1) Cov(X1 X2) . This formula hold for all 2· n· N. When n=N Var(SN)=0 and -- the sample is the entire population drawn out in random order. However, Cov(X_1,X_2) should not depend on the ultimate sample size, so we use the formula with n=N and obtain: Cov(X1 X2) = -s2/(N-1). And hence Var(Sn) = s2 n(1- (n-1)/(N-1)).