Download presentation
Presentation is loading. Please wait.
Published byKerry Hill Modified over 9 years ago
1
Exam Exam starts two weeks from today
2
Amusing Statistics Use what you know about normal distributions to evaluate this finding: The study, published in Pediatrics, the journal of the American Academy of Pediatrics, found that among the 4,508 students in Grades 5-8 ハ who participated, 36 per cent reported excellent school performance, 38 per cent reported good performance, 20 per cent said they were average performers, and 7 per cent said they performed below average.
3
Review The Z-test is used to compare the mean of a sample to the mean of a population and
4
Review The Z-score is normally distributed
5
Review The Z-score is normally distributed Thus the probability of obtaining any given Z-score by random sampling is given by the Z table
6
Review We can likewise determine critical values for Z such that we would reject the null hypothesis if our computed Z- score exceeds these values –For alpha =.05: Zcrit (one-tailed) = 1.64 Zcrit (two-tailed) = 1.96
7
Confidence Intervals A related question you might ask: –Suppose you’ve measured a mean and computed a standard error of that mean –What is the range of values such that there is a 95% chance of the population mean falling within that range?
8
There is a 2.5% chance that the population mean is actually 1.96 standard errors more than the observed mean Confidence Intervals 95% 1.96 2.5% True mean?
9
There is a 2.5% chance that the population mean is actually 1.96 standard errors less than the observed mean Confidence Intervals 2.5% -1.96 95% True mean?
10
Thus there is a 95% chance that the true population mean falls within + or - 1.96 standard errors from a sample mean Confidence Intervals
11
Thus there is a 95% chance that the true population mean falls within + or - 1.96 standard errors from a sample mean Likewise, there is a 95% chance that the true population mean falls within + or - 1.96 standard deviations from a single measurement Confidence Intervals
12
This is called the 95% confidence interval…and it is very useful It works like significance bounds…if the 95% C.I. doesn’t include the mean of a population you’re comparing your sample to, then your sample is significantly different from that population Confidence Intervals
13
Consider an example: You measure the concentration of mercury in your backyard to be.009 mg/kg The concentration of mercury in the Earth’s crust is.007 mg/kg. Let’s pretend that, when measured at many sites around the globe, the standard deviation is known to be.002 mg/kg Confidence Intervals
14
The 95% confidence interval for this mercury measurement is Confidence Intervals
15
This interval includes.007 mg/kg which, it turns out, is the mean concentration found in the earth’s crust in general Thus you would conclude that your backyard isn’t artificially contaminated by mercury Confidence Intervals
16
Imagine you take 25 samples from around Alberta and you found: Confidence Intervals
17
Imagine you take 25 samples from around Alberta and you found:.009 +/- (1.96 x.0004) =.008216 to.009784 This interval doesn’t include the.007 mg/kg value for the earth’s crust so you would conclude that Alberta has an artificially elevated amount of mercury in the soil Confidence Intervals
18
Power we perform a Z-test and determine that the difference between the mean of our sample and the mean of the population is not due to chance with a p <.05
19
Power we perform a Z-test and determine that the difference between the mean of our sample and the mean of the population is not due to chance with a p <.05 we say that we have a significant result…
20
Power we perform a Z-test and determine that the difference between the mean of our sample and the mean of the population is not due to chance with a p <.05 we say that we have a significant result… but what if p is >.05?
21
Power What are the two reasons why p comes out greater than.05?
22
Power What are the two reasons why p comes out greater than.05? –Your experiment lacked Statistical Power and you made a Type II Error –The null hypothesis really is true
23
Power Two approaches: –The Hopelessly Jaded Grad Student Solution –The Wise and Well Adjusted Professor Procedure
24
Power 1. Hopelessly Jaded Grad Student Solution - conclude that your hypothesis was wrong and go directly to the grad student pub
25
Power - This is not the recommended course of action
26
Power 2. The Wise Professor Procedure - consider the several reasons why you might not have detected a significant effect
27
Power - recommended by wise professors the world over
28
Power Why might p be greater than.05 ? Recall that: and
29
Power Why might p be greater than.05 ? 1. Small effect size: –The effect doesn’t stand out from the variability in the data – You might be able to increase your effect size (e.g. with a larger dose or treatment) is quite close to the mean of the population
30
Power Why might p be greater than.05 ? 2. Noisy Data –A large denominator will swamp the small effect –Take greater care to reduce measurement errors and thereforeis quite large
31
Power Why might p be greater than.05 ? 3. Sample Size is Too Small –A large denominator will swamp the small effect –Run more subjects is quite large becauseis small
32
Power The solution in each case is more power:
33
Power The solution in each case is more power: Power is like sensitivity - the ability to detect small effects in noisy data
34
Power The solution in each case is more power: Power is like sensitivity - the ability to detect small effects in noisy data It is the opposite of Type II Error rate
35
Power The solution in each case is more power: Power is like sensitivity - the ability to detect small effects in noisy data It is the opposite of Type II Error rate So that you know: there are equations for computing statistical power
36
Power An important point about power and the null hypothesis: –Failing to reject the null hypothesis DOES NOT PROVE it to be true!!!
37
Power Consider an example: –How to prove that smoking does not cause cancer: enroll 2 people who smoke infrequently and use an antique X-Ray camera to look for cancer Compare the mean cancer rate in your group (which will probably be zero) to the cancer rate in the population (which won’t be) with a Z-test
38
Power Consider an example: –If p came out greater than.05, you still wouldn’t believe that smoking doesn’t cause cancer
39
Power Consider an example: –If p came out greater than.05, you still wouldn’t believe that smoking doesn’t cause cancer –You will, however, often encounter statements such as “The study failed to find…” misinterpreted as “The study proved no effect of…”
40
We’ve been using examples in which a single sample is compared to a population Experimental Design
41
We’ve been using examples in which a single sample is compared to a population Often we employ more sophisticated designs Experimental Design
42
We’ve been using examples in which a single sample is compared to a population Often we employ more sophisticated designs What are some different ways you could run an experiment? Experimental Design
43
Compare one mean to some value –Often that value is zero
44
Experimental Design Compare one mean to some value –Often that value is zero Compare two means to each other
45
Experimental Design There are two general categories of comparing two (or more) means with each other
46
Experimental Design 1.Repeated Measures - also called “within-subjects” comparison The same subjects are given pre- and post- measurements e.g. before and after taking a drug to lower blood pressure Powerful because variability between subjects is factored out Note that pre- and post- scores are linked - we say that they are dependant Note also that you could have multiple tests
47
Experimental Design 1.Problems with Repeated-Measure design: Practice/Temporal effect - subjects get better/worse over time The act of measuring might preclude further measurement - e.g. measuring brain size via surgery Practice effect - subjects improve with repeated exposure to a procedure
48
Experimental Design 2. Between-Subjects Design Subjects are randomly assigned to treatment groups - e.g. drug and placebo Measurements are assumed to be statistically independent
49
Experimental Design 2. Problems with Between-Subjects design Can be less powerful because variability between two groups of different subjects can look like a treatment effect Often needs more subjects
50
Experimental Design We’ll need some statistical tests that can compare: –One sample mean to a fixed value –Two dependent sample means to each other (within-subject) –Two independent sample means to each other (between-subject)
51
Experimental Design The t-test can perform each of these functions It also gets around a big problem with the z-test…
52
Problems with Z and what to do instead
53
The Z statistic The Z statistic (with which to compare to the Zcrit) Where
54
The Z statistic What is the problem you will encounter in trying to use this statistic?
55
The Z statistic What is the problem you will encounter in trying to use this statistic? Although you might have a guess about the population mean, you will almost certainly not know the population variance!
56
The Z statistic Where
57
The Z statistic Where
58
The Z statistic Where
59
The Z statistic Where
60
The Z statistic What to do? Could we estimate What would we use and what would have to be the case for it to be useful?
61
The Z statistic What to do? Could we estimate What would we use and what would have to be the case for it to be useful? We could use our sample variance, S 2 to estimate the population variance
62
Estimating Population Variance Just like there are many sample means (the sampling distribution of the mean) there are many S 2 s
63
Estimating Population Variance Just like there are many sample means (the sampling distribution of the mean) there are many S 2 s tends to be near the value of but does S 2 tend to be near the value of
64
Estimating Population Variance Just like there are many sample means (the sampling distribution of the mean) there are many S 2 s tends to be near the value of but does S 2 tend to be near the value of No. It is a biased estimator. It tends to be lower than
65
Estimating Population Variance Why is S 2 biased?
66
Estimating Population Variance Why is S 2 biased? The sum of the deviation scores in your sample must equal zero regardless of where they came from in the population
67
Estimating Population Variance Why is S 2 biased? The sum of the deviation scores in your sample must equal zero regardless of where they came from in the population This means that the deviations in your sample are somewhat more constrained than in the population
68
Estimating Population Variance Why is S 2 biased? The sum of the deviation scores in your sample must equal zero regardless of where they came from in the population This means that the deviations in your sample are somewhat more constrained than in the population S 2 is has relatively fewer degrees of freedom than the entire population
69
Estimating Population Variance Specifically S 2 has n - 1 degrees of freedom
70
Estimating Population Variance Specifically S 2 has n - 1 degrees of freedom So if we compute S 2 but use n - 1 instead of n in the denominator we’ll get an unbiased estimator of
71
Estimating Population Variance Of course if you’ve already computed S 2 using n in the denominator you can multiply by n to recover the sum of squared deviations and then divide by n-1
72
The t Statistic(s) Using an estimated, which we’ll call we can create an estimate of which we’ll call where
73
The t Statistic(s) Using, instead of we get a statistic that isn’t from a normal (Z) distribution - it is from a family of distributions called t
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.