Topic 5 - Joint distributions and the CLT

Slides:



Advertisements
Similar presentations
Random Processes Introduction (2)
Advertisements

Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
Distributions of sampling statistics Chapter 6 Sample mean & sample variance.
Chapter 2 Multivariate Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Joint Probability Distributions and Random Samples
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Sampling Distributions (§ )
THE CENTRAL LIMIT THEOREM The “World is Normal” Theorem.
Copyright © 2006 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Econ 140 Lecture 61 Inference about a Mean Lecture 6.
Use of moment generating functions. Definition Let X denote a random variable with probability density function f(x) if continuous (probability mass function.
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
Slide 9- 1 Copyright © 2010 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Business Statistics First Edition.
Statistics Lecture 18. Will begin Chapter 5 today.
Probability Theory Part 2: Random Variables. Random Variables  The Notion of a Random Variable The outcome is not always a number Assign a numerical.
Sampling Distributions
Assignment 2 Chapter 2: Problems  Due: March 1, 2004 Exam 1 April 1, 2004 – 6:30-8:30 PM Exam 2 May 13, 2004 – 6:30-8:30 PM Makeup.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
4. Convergence of random variables  Convergence in probability  Convergence in distribution  Convergence in quadratic mean  Properties  The law of.
Standard Normal Distribution
Stat 321 – Lecture 19 Central Limit Theorem. Reminders HW 6 due tomorrow Exam solutions on-line Today’s office hours: 1-3pm Ch. 5 “reading guide” in Blackboard.
Chapter 7: Variation in repeated samples – Sampling distributions
A random variable that has the following pmf is said to be a binomial random variable with parameters n, p The Binomial random variable.
The moment generating function of random variable X is given by Moment generating function.
Continuous Random Variables and Probability Distributions
6-5 The Central Limit Theorem
Normal and Sampling Distributions A normal distribution is uniquely determined by its mean, , and variance,  2 The random variable Z = (X-  /  is.
Continuous Probability Distribution  A continuous random variables (RV) has infinitely many possible outcomes  Probability is conveyed for a range of.
Standard error of estimate & Confidence interval.
Sampling Distributions  A statistic is random in value … it changes from sample to sample.  The probability distribution of a statistic is called a sampling.
Pairs of Random Variables Random Process. Introduction  In this lecture you will study:  Joint pmf, cdf, and pdf  Joint moments  The degree of “correlation”
Jointly Distributed Random Variables
Copyright ©2011 Nelson Education Limited The Normal Probability Distribution CHAPTER 6.
AP Statistics 9.3 Sample Means.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5.
Chapter 10 – Sampling Distributions Math 22 Introductory Statistics.
Estimation This is our introduction to the field of inferential statistics. We already know why we want to study samples instead of entire populations,
Discrete distribution word problems –Probabilities: specific values, >, =, … –Means, variances Computing normal probabilities and “inverse” values: –Pr(X
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc.. Chap 7-1 Developing a Sampling Distribution Assume there is a population … Population size N=4.
Population and Sample The entire group of individuals that we want information about is called population. A sample is a part of the population that we.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
8 Sampling Distribution of the Mean Chapter8 p Sampling Distributions Population mean and standard deviation,  and   unknown Maximal Likelihood.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Maz Jamilah Masnan Institute of Engineering Mathematics Semester I 2015/ Sampling Distribution of Mean and Proportion EQT271 ENGINEERING STATISTICS.
Chapter 8 Sampling Variability and Sampling Distributions.
1 Topic 5 - Joint distributions and the CLT Joint distributions –Calculation of probabilities, mean and variance –Expectations of functions based on joint.
Sampling Distributions Sampling Distribution of Sampling Distribution of Point Estimation Point Estimation Introduction Introduction Sampling Distribution.
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
1 Probability and Statistical Inference (9th Edition) Chapter 4 Bivariate Distributions November 4, 2015.
Continuous Random Variables and Probability Distributions
1 Sampling distributions The probability distribution of a statistic is called a sampling distribution. : the sampling distribution of the mean.
Chapter 9 Inferences Based on Two Samples: Confidence Intervals and Tests of Hypothesis.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a number that describes the population. In statistical practice, the value of.
Random Variables By: 1.
Sampling and Sampling Distributions
Ch5.4 Central Limit Theorem
Sampling Distribution Estimation Hypothesis Testing
Statistics Lecture 19.
STAT 311 REVIEW (Quick & Dirty)
The distribution function F(x)
Chapter 7: Sampling Distributions
Introduction to Probability & Statistics The Central Limit Theorem
Sampling Distributions (§ )
Mathematical Expectation
Presentation transcript:

Topic 5 - Joint distributions and the CLT Joint distributions - pages 145 - 156 Central Limit Theorem - pages 183 - 185

Joint distributions Often times, we are interested in more than one random variable at a time. For example, what is the probability that a car will have at least one engine problem and at least one blowout during the same week? X = # of engine problems in a week Y = # of blowouts in a week P(X ≥ 1, Y ≥ 1) is what we are looking for To understand these sorts of probabilities, we need to develop joint distributions.

Discrete distributions A discrete joint probability mass function is given by f(x,y) = P(X = x, Y = y) where

Return to the car example Consider the following joint pmf for X and Y P(X ≥ 1, Y ≥ 1) = P(X ≥ 1) = E(X + Y) = X\Y 1 2 3 4 1/2 1/16 1/32

Joint to marginals The probability mass functions for X and Y individually (called marginals) are given by Returning to the car example: fX(x) = fY(y) = E(X) = E(Y) =

Continuous distributions A joint probability density function for two continuous random variables, (X,Y), has the following four properties:

Continuous example Consider the following joint pdf: Show condition 2 holds on your own. Show P(0 < X < 1, ¼ < Y < ½) = 23/512

Joint to marginals The marginal pdfs for X and Y can be found by For the previous example, find fX(x) and fY(y).

Independence of X and Y The random variables X and Y are independent if f(x,y) = fX(x) fY(y) for all pairs (x,y). For the discrete clunker car example, are X and Y independent? For the continuous example, are X and Y independent?

Sampling distributions We assume that each data value we collect represents a random selection from a common population distribution. The collection of these independent random variables is called a random sample from the distribution. A statistic is a function of these random variables that is used to estimate some characteristic of the population distribution. The distribution of a statistic is called a sampling distribution. The sampling distribution is a key component to making inferences about the population.

StatCrunch example StatCrunch subscriptions are sold for 6 months ($5) or 12 months ($8). From past data, I can tell you that roughly 80% of subscriptions are $5 and 20% are $8. Let X represent the amount in $ of a purchase. E(X) = Var(X) =

StatCrunch example continued Now consider the amounts of a random sample of two purchases, X1, X2. A natural statistic of interest is X1 + X2, the total amount of the purchases. Outcomes X1 + X2 Probability X1 + X2 Probability

StatCrunch example continued E(X1 + X2) = E([X1 + X2]2) = Var(X1 + X2) =

StatCrunch example continued If I have n purchases in a day, what is my expected earnings? the variance of my earnings? the shape of my earnings distribution for large n? Let’s experiment by simulating 1000 days with 100 purchases per day. StatCrunch These are notes for this page.

Central Limit Theorem We have just illustrated one of the most important theorems in statistics. As the sample size, n, becomes large the distribution of the sum of a random sample from a distribution with mean m and variance s2 converges to a Normal distribution with mean nm and variance ns2. A sample size of at least 30 is typically required to use the CLT The amazing part of this theorem is that it is true regardless of the form of the underlying distribution.

Airplane example Suppose the weight of an airline passenger has a mean of 150 lbs. and a standard deviation of 25 lbs. What is the probability the combined weight of 100 passengers will exceed the maximum allowable weight of 15,500 lbs? How many passengers should be allowed on the plane if we want this probability to be at most 0.01?

The sample mean For constant c, E(cY) = cE(Y) and Var(cY) = c2Var(Y) The CLT says that for large samples, is approximately normal with a mean of m and a variance of s2/n. So, the variance of the sample mean decreases with n.

Sampling applet