Central Limit Theorem Example: (NOTE THAT THE ANSWER IS CORRECTED COMPARED TO NOTES5.PPT) –5 chemists independently synthesize a compound 1 time each.

Slides:



Advertisements
Similar presentations
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Advertisements

Statistics: Purpose, Approach, Method. The Basic Approach The basic principle behind the use of statistical tests of significance can be stated as: Compare.
Sampling Distributions (§ )
THE CENTRAL LIMIT THEOREM The “World is Normal” Theorem.
Copyright © 2009 Cengage Learning 9.1 Chapter 9 Sampling Distributions.
McGraw-Hill Ryerson Copyright © 2011 McGraw-Hill Ryerson Limited. Adapted by Peter Au, George Brown College.
Sampling distributions. Example Take random sample of students. Ask “how many courses did you study for this past weekend?” Calculate a statistic, say,
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Point and Confidence Interval Estimation of a Population Proportion, p
Hypothesis Testing: Type II Error and Power.
Sampling Distributions
BCOR 1020 Business Statistics
Sampling Theory Determining the distribution of Sample statistics.
R. Kass/S07 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
1 Sampling Distributions Presentation 2 Sampling Distribution of sample proportions Sampling Distribution of sample means.
Sociology 5811: Lecture 7: Samples, Populations, The Sampling Distribution Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Inference for a Single Population Proportion (p).
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
The Standardized Normal Distribution Z is N( 0, 1 2 ) The standardized normal X is N( ,  2 ) 1.For comparison of several different normal distributions.
Sampling Distributions of Proportions. Parameter A number that describes the population Symbols we will use for parameters include  - mean  – standard.
Statistics for Data Miners: Part I (continued) S.T. Balke.
Stat 13, Tue 5/8/ Collect HW Central limit theorem. 3. CLT for 0-1 events. 4. Examples. 5.  versus  /√n. 6. Assumptions. Read ch. 5 and 6.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5.
Exam 1 is two weeks from today (March 9 th ) in class 15% of your grade Covers chapters 1-6 and the central limit theorem. I will put practice problems,
Discrete distribution word problems –Probabilities: specific values, >, =, … –Means, variances Computing normal probabilities and “inverse” values: –Pr(X
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
On Thursday, I’ll provide information about the project Due on Friday after last class. Proposal will be due two weeks from today (April 15 th ) You’re.
Introduction  Populations are described by their probability distributions and parameters. For quantitative populations, the location and shape are described.
Distributions of the Sample Mean
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Sample Variability Consider the small population of integers {0, 2, 4, 6, 8} It is clear that the mean, μ = 4. Suppose we did not know the population mean.
February 2012 Sampling Distribution Models. Drawing Normal Models For cars on I-10 between Kerrville and Junction, it is estimated that 80% are speeding.
Sampling Distribution Models Chapter 18. Toss a penny 20 times and record the number of heads. Calculate the proportion of heads & mark it on the dot.
Exam 1 next Thursday (March 7 th ) in class 15% of your grade Covers chapters 1-6 and the central limit theorem I will put practice problems, old exams,
Random Sampling Approximations of E(X), p.m.f, and p.d.f.
HW solutions are on the web. See website for how to calculate probabilities with minitab, excel, and TI calculators.
More Examples: There are 4 security checkpoints. The probability of being searched at any one is 0.2. You may be searched more than once in total and all.
Machine Learning Chapter 5. Evaluating Hypotheses
Review: Large Sample Confidence Intervals 1-  confidence interval for a mean: x +/- z  /2 s/sqrt(n) 1-  confidence interval for a proportion: p +/-
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a measure of the population. This value is typically unknown. (µ, σ, and now.
Chapter 9 Sampling Distributions Sir Naseer Shahzada.
Random Variables Numerical Quantities whose values are determine by the outcome of a random experiment.
8.1 Estimating µ with large samples Large sample: n > 30 Error of estimate – the magnitude of the difference between the point estimate and the true parameter.
Chapter 5 Probability Distributions 5-1 Overview 5-2 Random Variables 5-3 Binomial Probability Distributions 5-4 Mean, Variance and Standard Deviation.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
AP STATS: WARM UP I think that we might need a bit more work with graphing a SAMPLING DISTRIBUTION. 1.) Roll your dice twice (i.e. the sample size is 2,
Estimating a Population Proportion ADM 2304 – Winter 2012 ©Tony Quon.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a number that describes the population. In statistical practice, the value of.
Class Six Turn In: Chapter 15: 30, 32, 38, 44, 48, 50 Chapter 17: 28, 38, 44 For Class Seven: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 Read.
R. Kass/W04 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability.
4. Overview of Probability Network Performance and Quality of Service.
Sampling Distributions
THE CENTRAL LIMIT THEOREM
Keller: Stats for Mgmt & Econ, 7th Ed Sampling Distributions
Announcements Exams are graded. I’ll hand them back at end of class. Solution will be on the web. Many people did better than I think they’d guess… Easy.
Chapter 6: Sampling Distributions
The Gaussian Probability Distribution Function
Combining Random Variables
Chapter 7 Sampling Distributions.
Welcome Back Please hand in your homework.
Chapter 7 Sampling Distributions.
Chapter 7 Sampling Distributions.
Sampling Distributions (§ )
Chapter 7 Sampling Distributions.
Presentation transcript:

Central Limit Theorem Example: (NOTE THAT THE ANSWER IS CORRECTED COMPARED TO NOTES5.PPT) –5 chemists independently synthesize a compound 1 time each. –Each reaction should produce 10ml of a substance. –Historically, the amount produced by each reaction has been normally distributed with std dev 0.5ml. 1.What’s the probability that less than 49.8mls of the substance are made in total? 2.What’s the probability that the average amount produced is more than 10.1ml? 3.Suppose the average amount produced is more than 11.0ml. Is that a rare event? Why or why not? If more than 11.0ml are made, what might that suggest?

Answer: Central limit theorem: If E(X i )=  and Var(X i )=  2 for all i (and independent) then: X 1 +…+X n ~ N(n ,n  2 ) (X 1 +…+X n )/n ~ N( ,  2 /n)

Lab: 1.Let Y = total amount made. Y~N(5*10,5*0.5 2 ) (by CLT) Pr(Y<49.8) = Pr[(Y-50)/1.12 < ( )/1.12] =Pr(Z < -0.18) = Let W = average amount made. W~N(10,0.5 2 /5) (by CLT) Pr(W > 10.1) = Pr[Z > (10.1 – 10)/0.22] =Pr(Z > 0.45) = 0.33

Lab (continued) 3.One definition of rare: It’s a rare event if Pr(W > 11.0) is small (i.e. if “Seeing probability of 11.0 or something more extreme is small”) Pr(W>11) = Pr[Z > (11-10)/0.22] = Pr(Z>4.55) = approximately zero. This suggests that perhaps either the true mean is not 10 or true std dev is not 0.5 (or not normally distributed…)

Sample size: 1006 (source: gallup.com)

Let X i = 1 if person i thinks the President is hiding something and 0 otherwise. Suppose E(X i ) = p and Var(X i ) = p(1-p) and each person’s opinion is independent. Let Y = total number of “yesses” = X 1 +…+ X 1006 Y ~ Bin(1006,p) Suppose p = 0.36 (this is the estimate…) What is Pr(Y < 352)? Note that this definition turns three outcomes into two outcomes

Normal Approximation to the binomial CDF –Even with computers, as n gets large, computing things like this can become difficult. (1006 is OK, but how about 1,000,000?) –Idea: Use the central limit theorem approximate this probability –Y is approximately N[1006*0.36, (0.36)*(0.64)*1006] = N(362.16,231.8) (by central limit theorem) Pr[ (Y )/15.2 < ( )/15.2] = Pr(Z < -0.67) = 0.25 Pr(Y<352) = Pr(Y=0)+…+Pr(Y=351), where Pr(Y=k) = (1006 choose k)0.36 k k

Normal Approximation to the binomial CDF Black “step function” is plots of bin(1006,0.36) pdf versus Y (integers) Blue line is plot of Normal(362.16,231.8) pdf

Normal Approximation to the binomial CDF Area under blue curve to left of 352 is approximately equal to the sum of areas of rectangles (black Stepfunction) to the left of 352

Comments about normal approximation of the binomial : Rule of thumb is that it’s OK if np>5 and n(1-p)>5. “Continuity correction” Y is binomial. If we use the normal approximation to the probability that Y<k, we should calculate Pr(Y<k+.5) If we use the normal approximation to the probability that Y>k, we should calculate Pr(Y<k-.5) (see picture on board)

Probability meaning of 6 sigma Even if you shift the process mean for the center of the specifications to 1.5 standard deviations toward one of the specifications, then you will expect no more than 3.4 out of a million defects outside of the specification toward which you shifted. (I know it’s convoluted, but that’s the definition…)

What does 6 sigma mean? (example) Suppose a product has a quantitative specification: ex: “Make the gap between the car door and the car body between 3.4 and 4.6mm.” When cars are actually made, the “std dev of car door gap is 0.1mm”. i.e. X 1,…,X n are gap widths. The sqrt(sample variance of X 1,…,X n )= 0.1mm

Lower specification Upper specification 3.4mm4.6mm 4.6 – 3.4 = 1.2 = 12*0.1 = 12*sigma Statistically, six sigma means that Upper Spec – Lower Spec > 12 sigma (i.e. Specs are fixed. Lower the manufactuing process variability.) Center of spec = 4mm gap Shifted mean = 3.85mm gap Distribution of gap widths Probability of being out here is Pr( gap is less than 3.4 ) = Pr( (gap – 3.85)/0.1 < ( )/.1) =Pr( Z < -4.5) = 3.4/1,000,000 Arbitrary “magic” number for 6 

In general: Assume process mean is 1.5 standard deviations toward the lower spec: i.e. E(X)=4-1.5  and assume X has a normal distribution. When the process is in control enough so that the distance between the center of the specs and the lower spec is least 6 , then Pr(X below lower spec) =Pr( X<4- 6  ) =Pr[(X- (4-1.5  -6  (4-1.5  ] =Pr(Z<-4.5) = 3.4/1,000,000 Probability meaning of 6 sigma

Control Charts Let X = an average of n measurements. Each measurement has mean  and variance  2. Fact: –By the central limit theorem, almost all observations of X fall in the interval  +/- 3  /sqrt(n) (i.e. mean +/- 3 standard deviations) –  /sqrt(n) is also called  x or standard error

Use the “fact” to detect changes in production quality Idea: let x i = average door gap from the n cars made by shift i at the car plant   +3  /sqrt(n) (Upper Control Limit)  -3  /sqrt(n) (Lower Control Limit) shift x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 x8x8 Points outside the +/- 3 std error bounds, are called “out of control”. They are evidence that  and or  are not the true mean and std dev any more, and the process needs to be readjusted. Calculate the “false alarm rate”… (= 26/10,000)

Assume 100 new people are polled. Assume true pr( a new person says yes) = Let P = “P hat” = number say yes/100 What’s an approximation to the distribution of P-hat? Use the approximation to determine a number so that the Pr(p-hat> that number) = 0.95.

EXAMPLE OF SAMPLING DISTRIBUTION OF P-HAT X k = 1 if person k says yes and 0 if not. Note that E(X k )=0.36=p and Var(X k )=0.36*0.64=p(1-p) Note that X k is binomial(1,0.36). P-hat = (X 1 +…+X 100 )/100. By CLT, P-hat is approximately N(0.36,0.36*0.64/100). (Rule of thumb is that this approximation is good if np>5 and n(1-p)>5.)

Suppose true p is If survey is conducted again on 100 people, then P-hat ~ N(.36,(.36)(.64)/100) = N(.36, ) Want p 0 so that Pr(P-hat<p 0 ) = 0.95 Pr(P-hat<p 0 ) = 0.95 means Pr(Z < (p )/0.048) = Since Pr(Z<1.645) = 0.95, (p )/0.048 = (p ) = p 0 =

Suppose true p is If survey is conducted again on 49 people, what’s the probability of seeing 38% to 44% favorable responses? Pr( 0.38 < P ”hat” < 0.44) = Pr[( )/sqrt(0.40*0.60/49) < Z < ( )/sqrt(0.40*0.60/49) ] = Pr(-0.29 < Z < 0.57) = Pr(Z<0.57) – Pr(Z<-0.29) = =0.3298