Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.

Slides:



Advertisements
Similar presentations
Joint Probability Distributions and Random Samples
Advertisements

1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Sampling Variability & Sampling Distributions.
Chapter 7 Introduction to Sampling Distributions
Sampling Distributions
Evaluating Hypotheses
Definitions Uniform Distribution is a probability distribution in which the continuous random variable values are spread evenly over the range of possibilities;
Sampling Distributions
Introduction to Probability and Statistics Chapter 7 Sampling Distributions.
Chapter 7 The Normal Probability Distribution 7.5 Sampling Distributions; The Central Limit Theorem.
Chapter 7 Probability and Samples: The Distribution of Sample Means
The Lognormal Distribution
Sample Distribution Models for Means and Proportions
Normal and Sampling Distributions A normal distribution is uniquely determined by its mean, , and variance,  2 The random variable Z = (X-  /  is.
Continuous Probability Distribution  A continuous random variables (RV) has infinitely many possible outcomes  Probability is conveyed for a range of.
Joint Probability Distributions and Random Samples
© Copyright McGraw-Hill CHAPTER 6 The Normal Distribution.
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 8 Continuous.
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
Copyright © 1998, Triola, Elementary Statistics Addison Wesley Longman 1 Normal Distribution as an Approximation to the Binomial Distribution Section 5-6.
Continuous Random Variables
Copyright ©2011 Nelson Education Limited The Normal Probability Distribution CHAPTER 6.
1 Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Chapter 6. Continuous Random Variables Reminder: Continuous random variable.
Copyright © Cengage Learning. All rights reserved. 7 Sample Variability.
Lecture 15: Statistics and Their Distributions, Central Limit Theorem
The Normal Distribution Chapter 6. Outline 6-1Introduction 6-2Properties of a Normal Distribution 6-3The Standard Normal Distribution 6-4Applications.
ENGR 610 Applied Statistics Fall Week 3 Marshall University CITE Jack Smith.
Bernoulli Trials Two Possible Outcomes –Success, with probability p –Failure, with probability q = 1  p Trials are independent.
The Practice of Statistics Third Edition Chapter 8: The Binomial and Geometric Distributions 8.1 The Binomial Distribution Copyright © 2008 by W. H. Freeman.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 7 Sampling Distributions.
Chapter 7 Probability and Samples: The Distribution of Sample Means.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
1 Chapter 8 Sampling Distributions of a Sample Mean Section 2.
Chapter 7 Sampling Distributions Statistics for Business (Env) 1.
Section 6-5 The Central Limit Theorem. THE CENTRAL LIMIT THEOREM Given: 1.The random variable x has a distribution (which may or may not be normal) with.
Chapter 8 Sampling Variability and Sampling Distributions.
June 11, 2008Stat Lecture 10 - Review1 Midterm review Chapters 1-5 Statistics Lecture 10.
Chapter 7: Sampling Distributions Section 7.1 How Likely Are the Possible Values of a Statistic? The Sampling Distribution.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Copyright © Cengage Learning. All rights reserved. 3 Discrete Random Variables and Probability Distributions.
Chapter 5 Sampling Distributions. Introduction Distribution of a Sample Statistic: The probability distribution of a sample statistic obtained from a.
Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.
1 Two Discrete Random Variables The probability mass function (pmf) of a single discrete rv X specifies how much probability mass is placed on each possible.
Probability & Statistics Review I 1. Normal Distribution 2. Sampling Distribution 3. Inference - Confidence Interval.
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
Chapter 6 Continuous Random Variables Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
7 Normal Curves and Sampling Distributions
The normal distribution
Sampling and Sampling Distributions
Ch5.4 Central Limit Theorem
Central Limit Theorem Section 5-5
Chapter 7 Review.
Statistical Intervals Based on a Single Sample
Sampling Variability & Sampling Distributions
Joint Probability Distributions and Random Samples
Sampling Distributions and Estimation
Chapter 6. Continuous Random Variables
Distribution of the Sample Means
6 Normal Curves and Sampling Distributions
Chapter 7 ENGR 201: Statistics for Engineers
Two Discrete Random Variables
Combining Random Variables
MATH 2311 Section 4.4.
Using the Tables for the standard normal distribution
The normal distribution
8.1 The Binomial Distribution
Sampling Distributions
6 Normal Curves and Sampling Distributions
MATH 2311 Section 4.4.
Presentation transcript:

Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples

Copyright © Cengage Learning. All rights reserved. 5.4 The Distribution of the Sample Mean

3 The importance of the sample mean springs from its use in drawing conclusions about the population mean . Some of the most frequently used inferential procedures are based on properties of the sampling distribution of. A preview of these properties appeared in the calculations and simulation experiments of the previous section, where we noted relationships between E( ) and  and also among V( ),  2, and n.

4 The Distribution of the Sample Mean Proposition Let X 1, X 2,..., X n be a random sample from a distribution with mean value  and standard deviation . Then In addition, with T 0 = X X n (the sample total),

5 The Distribution of the Sample Mean According to Result 1, the sampling (i.e., probability) distribution of is centered precisely at the mean of the population from which the sample has been selected. Result 2 shows that the distribution becomes more concentrated about  as the sample size n increases. In marked contrast, the distribution of T o becomes more spread out as n increases. Averaging moves probability in toward the middle, whereas totaling spreads probability out over a wider and wider range of values.

6 The Distribution of the Sample Mean The standard deviation is often called the standard error of the mean; it describes the magnitude of a typical or representative deviation of the sample mean from the population mean.

7 Example 24 In a notched tensile fatigue test on a titanium specimen, the expected number of cycles to first acoustic emission (used to indicate crack initiation) is  = 28,000, and the standard deviation of the number of cycles is  = Let X 1, X 2,..., X 25 be a random sample of size 25, where each X i is the number of cycles on a different randomly selected specimen. Then the expected value of the sample mean number of cycles until first emission is E( )  = 28,000, and the expected total number of cycles for the 25 specimens is E(T o ) = n  = 25(28,000) = 700,000.

8 Example 24 The standard deviation of (standard error of the mean) and of T o are If the sample size increases to n = 100, E( ) is unchanged, but = 500, half of its previous value (the sample size must be quadrupled to halve the standard deviation of ). cont’d

9 The Case of a Normal Population Distribution

10 The Case of a Normal Population Distribution Proposition Let X 1, X 2,..., X n be a random sample from a normal distribution with mean  and standard deviation . Then for any n, is normally distributed (with mean  and standard deviation, as is T o (with mean n  and standard Deviation ). We know everything there is to know about the and T o distributions when the population distribution is normal. In particular, probabilities such as P(a   b) and P(c  T o  d) can be obtained simply by standardizing.

11 The Case of a Normal Population Distribution Figure 5.14 illustrates the proposition. A normal population distribution and sampling distributions Figure 5.14

12 Example 25 The time that it takes a randomly selected rat of a certain subspecies to find its way through a maze is a normally distributed rv with  = 1.5 min and  =.35 min. Suppose five rats are selected. Let X 1,..., X 5 denote their times in the maze. Assuming the X i ’s to be a random sample from this normal distribution, what is the probability that the total time T o = X X 5 for the five is between 6 and 8 min?

13 Example 25 By the proposition, T o has a normal distribution with = n  = 5(1.5) = 7.5 and variance = n  2 = 5(.1225) =.6125, so =.783. To standardize T o, subtract and divide by : cont’d

14 Example 25 Determination of the probability that the sample average time (a normally distributed variable) is at most 2.0 min requires =  = 1.5 and = = Then cont’d

15 The Central Limit Theorem

16 The Central Limit Theorem When the X i ’s are normally distributed, so is for every sample size n. Even when the population distribution is highly nonnormal, averaging produces a distribution more bell-shaped than the one being sampled. A reasonable conjecture is that if n is large, a suitable normal curve will approximate the actual distribution of. The formal statement of this result is the most important theorem of probability.

17 The Central Limit Theorem Theorem The Central Limit Theorem (CLT) Let X 1, X 2,..., X n be a random sample from a distribution with mean  and variance  2. Then if n is sufficiently large, has approximately a normal distribution with and and T o also has approximately a normal distribution with The larger the value of n, the better the approximation.

18 The Central Limit Theorem Figure 5.15 illustrates the Central Limit Theorem. The Central Limit Theorem illustrated Figure 5.15

19 The Central Limit Theorem According to the CLT, when n is large and we wish to calculate a probability such as P(a   b), we need only “pretend” that is normal, standardize it, and use the normal table. The resulting answer will be approximately correct. The exact answer could be obtained only by first finding the distribution of, so the CLT provides a truly impressive shortcut.

20 Example 26 The amount of a particular impurity in a batch of a certain chemical product is a random variable with mean value 4.0 g and standard deviation 1.5 g. If 50 batches are independently prepared, what is the (approximate) probability that the sample average amount of impurity is between 3.5 and 3.8 g? According to the rule of thumb to be stated shortly, n = 50 is large enough for the CLT to be applicable.

21 Example 26 then has approximately a normal distribution with mean value = 4.0 and so cont’d

22 The Central Limit Theorem The CLT provides insight into why many random variables have probability distributions that are approximately normal. For example, the measurement error in a scientific experiment can be thought of as the sum of a number of underlying perturbations and errors of small magnitude. A practical difficulty in applying the CLT is in knowing when n is sufficiently large. The problem is that the accuracy of the approximation for a particular n depends on the shape of the original underlying distribution being sampled.

23 The Central Limit Theorem If the underlying distribution is close to a normal density curve, then the approximation will be good even for a small n, whereas if it is far from being normal, then a large n will be required. Rule of Thumb If n > 30, the Central Limit Theorem can be used. There are population distributions for which even an n of 40 or 50 does not suffice, but such distributions are rarely encountered in practice.

24 The Central Limit Theorem On the other hand, the rule of thumb is often conservative; for many population distributions, an n much less than 30 would suffice. For example, in the case of a uniform population distribution, the CLT gives a good approximation for n  12.

25 Other Applications of the Central Limit Theorem

26 Other Applications of the Central Limit Theorem The CLT can be used to justify the normal approximation to the binomial distribution discussed earlier. We know that a binomial variable X is the number of successes in a binomial experiment consisting of n independent success/failure trials with p = P(S) for any particular trial. Define a new rv X 1 by and define X 2, X 3,..., X n analogously for the other n – 1 trials. Each X i indicates whether or not there is a success on the corresponding trial.

27 Other Applications of the Central Limit Theorem Because the trials are independent and P(S) is constant from trial to trial, the X i ’s are iid (a random sample from a Bernoulli distribution). The CLT then implies that if n is sufficiently large, both the sum and the average of the X i ’s have approximately normal distributions.

28 Other Applications of the Central Limit Theorem When the X i ’s are summed, a 1 is added for every S that occurs and a 0 for every F, so X X n = X. The sample mean of the X i ’s is X/n, the sample proportion of successes. That is, both X and X/n are approximately normal when n is large.

29 Other Applications of the Central Limit Theorem The necessary sample size for this approximation depends on the value of p: When p is close to.5, the distribution of each X i is reasonably symmetric (see Figure 5.19), whereas the distribution is quite skewed when p is near 0 or 1. Using the approximation only if both np  10 and n(1  p)  10 ensures that n is large enough to overcome any skewness in the underlying Bernoulli distribution. Two Bernoulli distributions: (a) p =.4 (reasonably symmetric); (b) p =.1 (very skewed) Figure 5.19 (a) (b)

30 Other Applications of the Central Limit Theorem We know that X has a lognormal distribution if ln(X) has a normal distribution. Proposition Let X 1, X 2,..., X n be a random sample from a distribution for which only positive values are possible [P(X i > 0) = 1]. Then if n is sufficiently large, the product Y = X 1 X X n has approximately a lognormal distribution.

31 Other Applications of the Central Limit Theorem To verify this, note that Since ln(Y) is a sum of independent and identically distributed rv’s [the ln(X i )s], it is approximately normal when n is large, so Y itself has approximately a lognormal distribution.

32 Other Applications of the Central Limit Theorem As an example of the applicability of this result, Bury (Statistical Models in Applied Science,Wiley, p. 590) argues that the damage process in plastic flow and crack propagation is a multiplicative process, so that variables such as percentage elongation and rupture strength have approximately lognormal distributions.