Lectures prepared by: Elchanan Mossel elena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.

Slides:



Advertisements
Similar presentations
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Advertisements

Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Independence of random variables
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
The moment generating function of random variable X is given by Moment generating function.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Week 51 Theorem For g: R  R If X is a discrete random variable then If X is a continuous random variable Proof: We proof it for the discrete case. Let.
Chapter 16: Random Variables
Chris Morgan, MATH G160 February 3, 2012 Lecture 11 Chapter 5.3: Expectation (Mean) and Variance 1.
Review of Probability and Statistics
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 16 Random Variables.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
1 Lecture 4. 2 Random Variables (Discrete) Real-valued functions defined on a sample space are random vars. determined by outcome of experiment, we can.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Continuous Distributions The Uniform distribution from a to b.
1 Chapter 16 Random Variables. 2 Expected Value: Center A random variable assumes a value based on the outcome of a random event.  We use a capital letter,
X = 2*Bin(300,1/2) – 300 E[X] = 0 Y = 2*Bin(30,1/2) – 30 E[Y] = 0.
The Mean of a Discrete RV The mean of a RV is the average value the RV takes over the long-run. –The mean of a RV is analogous to the mean of a large population.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
STA347 - week 51 More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 16 Random Variables.
Probability Refresher. Events Events as possible outcomes of an experiment Events define the sample space (discrete or continuous) – Single throw of a.
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
Expectation. Let X denote a discrete random variable with probability function p(x) (probability density function f(x) if X is continuous) then the expected.
BUS304 – Probability Theory1 Probability Distribution  Random Variable:  A variable with random (unknown) value. Examples 1. Roll a die twice: Let x.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Week 121 Law of Large Numbers Toss a coin n times. Suppose X i ’s are Bernoulli random variables with p = ½ and E(X i ) = ½. The proportion of heads is.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section.
Review Know properties of Random Variables
Section 10.5 Let X be any random variable with (finite) mean  and (finite) variance  2. We shall assume X is a continuous type random variable with p.d.f.
The Variance of a Random Variable Lecture 35 Section Fri, Mar 26, 2004.
Mean and Standard Deviation Lecture 23 Section Fri, Mar 3, 2006.
Stat 35b: Introduction to Probability with Applications to Poker Outline for the day: 1.Uniform, normal, and exponential. 2.Exponential example. 3.Uniform.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Lectures prepared by: Elchanan Mossel Yelena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Sections.
Week 61 Poisson Processes Model for times of occurrences (“arrivals”) of rare phenomena where λ – average number of arrivals per time period. X – number.
Basic statistics Usman Roshan.
Chapter 16 Random Variables.
Lectures prepared by: Elchanan Mossel Yelena Shvets
Mean and Standard Deviation
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Week 8 Chapter 14. Random Variables.
Chapter 4: Mathematical Expectation:
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Probability Review for Financial Engineers
Lectures prepared by: Elchanan Mossel Yelena Shvets
Mean and Standard Deviation
Mean and Standard Deviation
Chebychev, Hoffding, Chernoff
Lectures prepared by: Elchanan Mossel Yelena Shvets
9. Limit Theorems.
AP Statistics Chapter 16 Notes.
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Lectures prepared by: Elchanan Mossel Yelena Shvets
Mathematical Expectation
Presentation transcript:

Lectures prepared by: Elchanan Mossel elena Shvets Introduction to probability Stat 134 FAll 2005 Berkeley Follows Jim Pitman’s book: Probability Section 3.3

Histo 1 X = 2*Bin(300,1/2) – 300 E[X] = 0

Histo 2 Y = 2*Bin(30,1/2) – 30 E[Y] = 0

Histo 3 Z = 4*Bin(10,1/4) – 10 E[Z] = 0

Histo 4 W = 0 E[W] = 0

A natural question: Is there a good parameter that allow to distinguish between these distributions? Is there a way to measure the spread?

Variance and Standard Deviation The variance of X, denoted by Var(X) is the mean squared deviation of X from its expected value  = E(X): Var(X) = E[(X-  ) 2 ]. The standard deviation of X, denoted by SD(X) is the square root of the variance of X:

Computational Formula for Variance Proof: E[ (X-  ) 2 ] = E[X 2 – 2  X +  2 ] E[ (X-  ) 2 ] = E[X 2 ] – 2  E[X] +  2 E[ (X-  ) 2 ] = E[X 2 ] – 2  2 +  2 E[ (X-  ) 2 ] = E[X 2 ] – E[X] 2 Claim:

Variance and SD For a general distribution Chebyshev inequality states that for every random variable X, X is expected to be close to E(X) give or take a few SD(X). Chebyshev Inequality: For every random variable X and for all k > 0: P(|X – E(X)| ¸ k SD(X)) · 1/k 2.

Properties of Variance and SD 1.Claim: Var(X) ¸ 0. Pf: Var(X) =  (x-  ) 2 P(X=x) ¸ 0 2.Claim: Var(X) = 0 iff P[X=  ] = 1.

Chebyshev’s Inequality P(|X – E(X)| ¸ k SD(X)) · 1/k 2 proof: Let  = E(X) and  = SD(X). Observe that |X–  | ¸ k , |X–  | 2 ¸ k 2  2. The RV |X–  | 2 is non-negative, so we can use Markov’s inequality: P(|X–  | 2 ¸ k 2  2 ) · E [|X–  | 2 ] / k 2  2 P(|X–  | 2 ¸ k 2  2 ) ·  2 / k 2  2 = 1/k 2.

Variance of Indicators Suppose I A is an indicator of an event A with probability p. Observe that I A 2 = I A. AcAc A I A =1= I A 2 I A =0= I A 2 E(I A 2 ) = E(I A ) = P(A) = p, so: Var(I A ) = E(I A 2 ) – E(I A ) 2 = p – p 2 = p(1-p).

Variance of a Sum of Independent Random Variables Claim: if X 1, X 2, …, X n are independent then: Var (X 1 +X 2 +…+X n ) = Var (X 1 )+ Var (X 2 )+…+ Var (X n ). Pf: Suffices to prove for 2 random variables. E[( X+Y – E(X+Y) ) 2 ] = E[( X-E(X) + Y–E(Y) ) 2 ] = E[( X-E(X)) E[(X-E(X)) (Y-E(Y))] + E(Y–E(Y) ) 2 ]= Var(X) + Var(Y) + 2 E[(X-E(X))] E[(Y-E(Y))] (mult.rule) = Var(X) + Var(Y) + 0

Variance and Mean under scaling and shifts Claim: SD(aX + b) = |a| SD(X) Proof: Var[aX+b] = E[(aX+b – a  –b) 2 ] = = E[a 2 (X-  ) 2 ] = a 2  2 Corollary: If a random variable X has E(X) =  and SD(X) =  > 0, then X * =(X-  )/  has E(X * ) =0 and SD(X * )=1.

Square Root Law Let X 1, X 2, …, X n be independent random variables with the same distribution as X, and let S n be their sum: S n =  i=1 n X i, and their average, then:

Weak Law of large numbers Then for every  > 0 : Thm: Let X 1, X 2, … be a sequence of independent random variables with the same distribution. Let  denote the common expected value  = E(X i ).

Weak Law of large numbers Now Chebyshev inequality gives us: Proof: Let  = E(X i ) and  = SD(X i ). Then from the square root law we have: For a fixed  right hand side tends to 0 as n tends to 1.

The Normal Approximation Let S n = X 1 + … + X n be the sum of independent random variables with the same distribution. Then for large n, the distribution of S n is approximately normal with mean E(S n ) = n  and SD(S n ) =  n 1/2, where  = E(X i ) and  = SD(X i ). In other words:

Sums of iid random variables Suppose X i represents the number obtained on the i’th roll of a die. Then X i has a uniform distribution on the set {1,2,3,4,5,6}.

Distribution of X 1

Sum of two dice We can obtain the distribution of S 2 = X 1 +X 2 by the convolution formula: P(S 2 = k) =  i=1 k-1 P(X 1 =i) P(X 2 =k-i| X 1 =i), by independence =  i=1 k-1 P(X 1 =i) P(X 2 =k-i).

Distribution of S 2

Sum of four dice We can obtain the distribution of S 4 = X 1 + X 2 + X 3 + X 4 = S 2 + S’ 2 again by the convolution formula: P(S 4 = k) =  i=1 k-1 P(S 2 =i) P(S’ 2 =k-i| S 2 =i), by independence of S 2 and S’ 2 =  i=1 k-1 P(S 2 =i) P(S’ 2 =k-i).

Distribution of S 4

Distribution of S 8

Distribution of S 16

Distribution of S 32

Distribution of X 1

Distribution of S 2

Distribution of S 4

Distribution of S 8

Distribution of S 16

Distribution of S 32

Distribution of X 1

Distribution of S 2

Distribution of S 4

Distribution of S 8

Distribution of S 16

Distribution of S 32