Stat 155, Section 2, Last Time Interpreted Midterm Results Variance of Random Variables –From Probability Table –Properties: Ignores shift Multiples comes.

Slides:



Advertisements
Similar presentations
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Advertisements

Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Sampling Distribution of a Sample Proportion Lecture 26 Sections 8.1 – 8.2 Wed, Mar 8, 2006.
Sampling Distributions
Copyright © 2009 Cengage Learning 9.1 Chapter 9 Sampling Distributions.
Copyright © 2010 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
McGraw-Hill Ryerson Copyright © 2011 McGraw-Hill Ryerson Limited. Adapted by Peter Au, George Brown College.
Chapter 7 Introduction to Sampling Distributions
Sampling Distributions
Part III: Inference Topic 6 Sampling and Sampling Distributions
Introduction to Probability and Statistics Chapter 7 Sampling Distributions.
1 The Sample Mean rule Recall we learned a variable could have a normal distribution? This was useful because then we could say approximately.
Binomial Probability Distribution.
Determining the Size of
Inferential Statistics
Fall 2013Biostat 5110 (Biostatistics 511) Week 7 Discussion Section Lisa Brown Medical Biometry I.
Chapter 5 Sampling Distributions
Introduction Parameters are numerical descriptive measures for populations. For the normal distribution, the location and shape are described by  and.
Sampling distributions BPS chapter 11 © 2006 W. H. Freeman and Company.
Last Time Q-Q plots –Q-Q Envelope to understand variation Applications of Normal Distribution –Population Modeling –Measurement Error Law of Averages –Part.
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
Copyright ©2011 Nelson Education Limited The Normal Probability Distribution CHAPTER 6.
Last Time Central Limit Theorem –Illustrations –How large n? –Normal Approximation to Binomial Statistical Inference –Estimate unknown parameters –Unbiasedness.
Lecture 3 Preview: Interval Estimates and the Central Limit Theorem Review Populations, Samples, Estimation Procedures, and the Estimate’s Probability.
MATB344 Applied Statistics Chapter 7 Sampling Distributions I.Sampling Plans and Experimental Designs II.Statistics and Sampling Distributions III.“The.
1 Psych 5500/6500 Standard Deviations, Standard Scores, and Areas Under the Normal Curve Fall, 2008.
Copyright © 2009 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
1 Chapter 18 Sampling Distribution Models. 2 Suppose we had a barrel of jelly beans … this barrel has 75% red jelly beans and 25% blue jelly beans.
Stor 155, Section 2, Last Time Prediction in Regression –Given new point X 0, predict Y 0 –Confidence interval for mean –Prediction Interval for value.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 7 Sampling Distributions.
Stat 155, Section 2, Last Time Binomial Distribution –Normal Approximation –Continuity Correction –Proportions (different scale from “counts”) Distribution.
Sampling Distribution Models Chapter 18. Toss a penny 20 times and record the number of heads. Calculate the proportion of heads & mark it on the dot.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 7 Sampling and Sampling Distributions.
Stat 31, Section 1, Last Time Time series plots Numerical Summaries of Data: –Center: Mean, Medial –Spread: Range, Variance, S.D., IQR 5 Number Summary.
Chapter 18: Sampling Distribution Models
Sample Means & Proportions
Stats Lunch: Day 3 The Basis of Hypothesis Testing w/ Parametric Statistics.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 6 Random Variables 6.3 Binomial and Geometric.
Administrative Matters Midterm II Results Take max of two midterm scores:
Chapter 9 Sampling Distributions Sir Naseer Shahzada.
Stat 31, Section 1, Last Time Big Rules of Probability –The not rule –The or rule –The and rule P{A & B} = P{A|B}P{B} = P{B|A}P{A} Bayes Rule (turn around.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
Introduction Sample surveys involve chance error. Here we will study how to find the likely size of the chance error in a percentage, for simple random.
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a number that describes the population. In statistical practice, the value of.
Chapter 6 Sampling and Sampling Distributions
THE NORMAL DISTRIBUTION
Sampling Distribution of the Sample Mean
Chapter 7: Sampling Distributions
Last Time Proportions Continuous Random Variables Probabilities
Chapter 7 Review.
Sampling Distribution Models
Chapter 7 Sampling and Sampling Distributions
CHAPTER 6 Random Variables
Stat 31, Section 1, Last Time Sampling Distributions
Chapter 18: Sampling Distribution Models
Sampling Distribution of a Sample Proportion
Do now: • Which diagram on the right illustrates which distribution?
Chapter 7: Sampling Distributions
Sampling Distributions
Chapter 7: Sampling Distributions
Chapter 7: Sampling Distributions
Chapter 7: Sampling Distributions
Chapter 7: Sampling Distributions
Sampling Distributions (§ )
Chapter 7: Sampling Distributions
Chapter 7: Sampling Distributions
Sample Proportions Section 9.2
Presentation transcript:

Stat 155, Section 2, Last Time Interpreted Midterm Results Variance of Random Variables –From Probability Table –Properties: Ignores shift Multiples comes through squared Sum when independent Sampling Distributions –Binomial Distribution Calculate Probs with BINOMDIST Mean & Standard Deviation

Reading In Textbook Approximate Reading for Today’s Material: Pages , Approximate Reading for Next Class: Pages ,

Binomial Distribution Normal Approximation to the Binomial Idea: Bi(n,p) prob. histo. curve So can approximate Binomial probs with normal areas

Normal Approx. to Binomial Before modern software, this was a critical issue, since Binomial Table (C in text) only goes to n = 20. Normal Approx made this possible… Now still useful, since BINONDIST conks out around n = 1000 (but political polls need n ~ ).

Normal Approx. to Binomial Visualization of Normal Approximation: Bi(100,0.3): Looks really good Bi(20,0.5): Chunky, approx. a little weak Bi(20,0.05): p too small for good approx. Bi(20,0.95): p too big for good approx.

Normal Approx. to Binomial When is Normal Approximation “acceptable”? Textbooks “Rule of Thumb” (there are others): Need: np >= 10 and n(1-p) >= 10 Relate to above examples: Bi(20,0.5): np = n(1-p) = 10, boundary case Bi(20,0.05): np = 4 < 10, poor approx. Bi(20,0.95): n(1-p) = 4 < 10, poor approx.

Normal Approx. to Binomial Nice Illustration: Publisher’s Website Statistical Applets Normal Approx to Binomial Shows how rule of thumb makes sense

Normal Approx. to Binomial HW:

Normal Approx. to Binomial HW: C18: In a political poll of 2000, 1010 will vote for A. To decide how “safe” it is to predict A will win: a.Calculate P{X >= 1010}, for X ~ Bi(2000,1/2) (0.327) (“could happen”, so not safe to predict) b.Recalculate, assuming 1050 will vote for A (0.0127) (now have stronger evidence, will build on this)

Normal Approx. to Binomial A refinement: Continuity Correction Idea: Binomial only takes on Integer Values Not really Continuous Better viewpoint (i.e. better approx.): Replace points by “± ½ bars”

Continuity Correction Graphically: Approx P{X=0} by area of bar above (-0.5,0.5), under curve. I.e. P{X = 0} = P{-0.5<X<0.5}

Continuity Correction More examples:

Continuity Correction Next time, consider replacing above e.g by something with number lines….

Continuity Correction Excel example: Part 3 For Bi(20,0.2): Continuity Correction much better About an order of magnitude in % error

Continuity Correction Notes: Gives substantial improvement for small n But doesn’t matter much for large n Class rules: –do when easy –Skip when it adds complications…

Continuity Correction HW C 19: For X ~ Bi(25,0.4) a.Check this is on the boundary of “OK” for using a Normal Approximation. Find: b.P{6 < X <= 11} c.P{6 <= X < 11} Using:

Continuity Correction Using: i.The naïve Normal approx. (0.607,0.607) ii.Normal approx. w/ cont. corr. (0.653,0.548) iii.The exact Binomial Distribution (0.659,0.556 )

Binomial Distribution Two Important “scales”: i.Counts: X ~ Bi(n,p) (done above) ii.Proportions (~ percentages):, on [0,1] scale often very natural, e.g. political polls

Binomial for Proportions Relationship betweens means and SDs: (“on average, expect” )

Binomial for Proportions Normal Approx for Proportions: (just uses above means and SDs)

Binomial for Proportions HW: 5.25 work with both BINOMDIST and Normal Approx. (BINOMDIST: (a) (b) )

And now for something completely different…. An interesting advertisement:

Section 5.2: Distrib’n of Sample Means Idea: Study Probability Structure of Based on Drawn independently From same distribution, Having Expected Value: And Standard Deviation:

Expected Value of Sample Mean How does relate to ? Sample mean “has the same mean” as the original data.

Variance of Sample Mean Study “spread” (i.e. quantify variation) of Variance of Sample mean “reduced by ”

S. D. of Sample Mean Since Standard Deviation is square root of Variance, Take square roots to get: S. D. of Sample mean “reduced by ”

Mean & S. D. of Sample Mean Summary: Averaging: 1.Gives same centerpoint 2.Reduces variation by factor of Called “Law of Averages, Part I”

Law of Averages, Part I Some consequences (worth noting): To “double accuracy”, need 4 times as much data. For “10 times the accuracy”, need 100 times as much data.

Law of Averages, Part I Nice Illustration: Publisher’s Website Statistical Applets Law of Large Numbers

Law of Averages, Part I HW: 5.29

Distribution of Sample Mean Now know center and spread, what about “shape of distribution”? Case 1: If are indep. CAN SHOW: (knew these, news is “mound shape”) Thus work with NORMDIST & NORMINV

Distribution of Sample Mean Case 2: If are “almost anything” STILL HAVE: “approximately”

Distribution of Sample Mean Remarks: Mathematics: in terms of Called “Law of Averages, Part II” Also called “Central Limit Theorem” Gives sense in which Normal Distribution is in the center Hence name “Normal” (ostentatious?)

Law of Averages, Part II More Remarks: Thus we will work with NORMDIST & NORMINV a lot, for averages This is why Normal Dist’n is good model for many different populations (any sum of small indep. Random pieces) Also explains Normal Approximation to the Binomial

Normal Approx. to Binomial Explained by Law of Averages. II, since: For X ~ Binomial (n,p) Can represent X as: Where: Thus X is an average (rescaled sum), so Law of Averages gives Normal Dist’n

Law of Averages, Part II Nice Java Demo: 1 Dice (think n = 1): Average Dist’n is flat 2 Dice (n = 1): Average Dist’n is triangle … 5 Dice (n = 5): Looks quite “mound shaped”

Law of Averages, Part II Another cool one: Create U shaped distribut’n with mouse Simul. samples of size 2: non-Normal Size n = 5: more normal Size n = 10 or 25: mound shaped

Law of Averages, Part II Textbook version: Publisher’s Website Statistical Applets Central Limit Theorem

Law of Averages, Part II Class Example: Shows: Even starting from non-normal shape, Averages become normal More so for more averaging SD smaller with more averaging ( )

Law of Averages, Part II HW: 5.35, 5.37, 5.39, 5.45

And now for something completely different…. A statistics professor was describing sampling theory to his class, explaining how a sample can be studied and used to generalize to a population. One of the students in the back of the room kept shaking his head.

And now for something completely different…. "What's the matter?" asked the professor. "I don't believe it," said the student, "why not study the whole population in the first place?" The professor continued explaining the ideas of random and representative samples. The student still shook his head.

And now for something completely different…. The professor launched into the mechanics of proportional stratified samples, randomized cluster sampling, the standard error of the mean, and the central limit theorem. The student remained unconvinced saying, "Too much theory, too risky, I couldn't trust just a few numbers in place of ALL of them."

And now for something completely different…. Attempting a more practical example, the professor then explained the scientific rigor and meticulous sample selection of the Nielsen television ratings which are used to determine how multiple millions of advertising dollars are spent. The student remained unimpressed saying, "You mean that just a sample of a few thousand can tell us exactly what over 250 MILLION people are doing?"

And now for something completely different…. Finally, the professor, somewhat disgruntled with the skepticism, replied, "Well, the next time you go to the campus clinic and they want to do a blood test...tell them that's not good enough... tell them to TAKE IT ALL!!" From: GARY C. RAMSEYER