LCS829 Wilfried Karmaus Introduction into the determination of the sample size for epidemiologic studies Objectives   Understand the concept   Learn.

Slides:



Advertisements
Similar presentations
Surveys Sample Size By R. Heberto Ghezzo Ph.D. R. Heberto Ghezzo Ph.D. Meakins-Christie laboratories McGill University - Montreal - Canada.
Advertisements

Comparing Two Proportions (p1 vs. p2)
Chapter 10: Hypothesis Testing
Chapter 19 Confidence Intervals for Proportions.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 19 Confidence Intervals for Proportions.
Confidence Intervals for Proportions
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
Point and Confidence Interval Estimation of a Population Proportion, p
Lecture 2: Thu, Jan 16 Hypothesis Testing – Introduction (Ch 11)
Chi-square Test of Independence
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
Sample Size Determination In the Context of Hypothesis Testing
Copyright © 2010 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
Inferences About Process Quality
Chapter 9 Hypothesis Testing.
Ch. 9 Fundamental of Hypothesis Testing
1 BA 555 Practical Business Analysis Review of Statistics Confidence Interval Estimation Hypothesis Testing Linear Regression Analysis Introduction Case.
Statistics for Managers Using Microsoft® Excel 5th Edition
Sample size calculations
Chapter 19: Confidence Intervals for Proportions
Tests of significance: The basics BPS chapter 15 © 2006 W.H. Freeman and Company.
Thomas Songer, PhD with acknowledgment to several slides provided by M Rahbar and Moataza Mahmoud Abdel Wahab Introduction to Research Methods In the Internet.
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 14 Analysis.
Chapter 10 Hypothesis Testing
STAT 3130 Statistical Methods I Session 2 One Way Analysis of Variance (ANOVA)
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Fundamentals of Hypothesis Testing: One-Sample Tests
14. Introduction to inference
Evidence-Based Medicine 4 More Knowledge and Skills for Critical Reading Karen E. Schetzina, MD, MPH.
CHP400: Community Health Program - lI Research Methodology. Data analysis Hypothesis testing Statistical Inference test t-test and 22 Test of Significance.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Inference for a Single Population Proportion (p).
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
Evidence-Based Medicine 3 More Knowledge and Skills for Critical Reading Karen E. Schetzina, MD, MPH.
Comparing Two Population Means
CHAPTER 16: Inference in Practice. Chapter 16 Concepts 2  Conditions for Inference in Practice  Cautions About Confidence Intervals  Cautions About.
Estimation of Statistical Parameters
Understanding the Variability of Your Data: Dependent Variable Two "Sources" of Variability in DV (Response Variable) –Independent (Predictor/Explanatory)
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Confidence intervals are one of the two most common types of statistical inference. Use a confidence interval when your goal is to estimate a population.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Chapter 8 Delving Into The Use of Inference 8.1 Estimating with Confidence 8.2 Use and Abuse of Tests.
1 Chapter 9 Hypothesis Testing. 2 Chapter Outline  Developing Null and Alternative Hypothesis  Type I and Type II Errors  Population Mean: Known 
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests Statistics.
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
© Copyright McGraw-Hill 2000
Chapter 20 Testing Hypothesis about proportions
Copyright © 2010 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
6.1 Inference for a Single Proportion  Statistical confidence  Confidence intervals  How confidence intervals behave.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Overview.
Issues concerning the interpretation of statistical significance tests.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
© Copyright McGraw-Hill 2004
Understanding Basic Statistics Fourth Edition By Brase and Brase Prepared by: Lynn Smith Gloucester County College Chapter Nine Hypothesis Testing.
SAMPLING DISTRIBUTION OF MEANS & PROPORTIONS. SAMPLING AND SAMPLING VARIATION Sample Knowledge of students No. of red blood cells in a person Length of.
Learning Objectives After this section, you should be able to: The Practice of Statistics, 5 th Edition1 DESCRIBE the shape, center, and spread of the.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
Significance Tests for Regression Analysis. A. Testing the Significance of Regression Models The first important significance test is for the regression.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Statistics 19 Confidence Intervals for Proportions.
Class Seven Turn In: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 For Class Eight: Chapter 20: 18, 20, 24 Chapter 22: 34, 36 Read Chapters 23 &
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Critical Appraisal Course for Emergency Medicine Trainees Module 2 Statistics.
Lecture Notes and Electronic Presentations, © 2013 Dr. Kelly Significance and Sample Size Refresher Harrison W. Kelly III, Ph.D. Lecture # 3.
Chapter Nine Hypothesis Testing.
More on Inference.
Sample size calculation
More on Inference.
Interpreting Epidemiologic Results.
Presentation transcript:

LCS829 Wilfried Karmaus Introduction into the determination of the sample size for epidemiologic studies Objectives   Understand the concept   Learn to apply statistical programs to determine the sample size   Learn basic steps to use SAS   Collaborate in small groups

Objectives for sample size estimations Level 6: Evaluation (critique, appraise, judge) Level 5: Synthesize (integrate, collaborate) (20%) Level 4: Analyze (hypothesis, structure) Level 3: Application (utilize, produce) (30%) Level 2: Comprehension (translate, discuss) Level 1: Knowledge (define, enumerate, recall) (50%)

Validity and Precision (1) Fundamental concern: avoidance and/or control of error Error = difference between true values and study results Accuracy = lack of error Precision = lack of random errorValidity = lack or control of systematic error

Validity and Precision (2) validity target estimator actual estimator results Precision

Random error = part of our experience that we cannot explain/predict How to increase precision? · increase the size of the study · increase the efficiency of the study

What can go wrong in planning the sample size? What can go wrong in planning the sample size? The truth about exposure is: Data of a study without a health risk with a health risk assess the exposure as: correct negative: 1-  beta error with a health risk without a health risk alpha error correct positive: 1 -  Statistical power = 1 - 

Alpha error Alpha is traditionally set to 5%, but if the risks involved in the 'exposed' group is high, then it can be set to 10%. Conversely, if the subject is not critical it could be set to 1%. Alpha = 10% means that with a little evidence the difference will be claimed significant. Alpha = 1% means that you need a very strong evidence to claim significance.

Beta error Beta is defined as 1 - power. The power of a test is a function of the difference between the groups. If the difference is Zero, the Power = Alpha. If the difference is very large, the Power = 1. That is why we must specify Beta (1-power) to be small when the difference in prevalence is small. If a study finds no difference, the power plays a major role in defining the maximum size that the true difference in the population may have. A power of 50% is a coin-flip chance to identify the true association as statistically significant.

1.What are the hypotheses? 2. 2.What are reasonable scenarios for ‘true’ parameter values?  Check other studies  prevalence of exposures (and confounders)  assumed differences  variance of the outcome variable Statistical power or guestimations of the required sample sizes

Sample Size for Surveys - Objectives of the Survey The objectives of the study can be: A. A.To estimate a parameter prevalence, an Odds Ratio, etc. B. B.To test an hypothesis by comparing two (or more ) groups. C. C.To test if a prevalence, Odds Ratio or Relative Risk is significant. For example between exposed and non exposed, a 'clean ' city versus a 'polluted' city, etc. For each objective the sample size needed has different requirements.

Control statistical power by: Sample size Strength of the association Error variance Control for confounders or take multiple predictors into account (Use SAS,  homework) Use formula/ programs  homework

Sample Size for Surveys - Estimation If the objective is an estimation (prevalence or incidence), just the value is not sufficient. It must be accompanied by a confidence interval (CI). A x% CI means: If another survey is taken with the same characteristics as this, then the point estimate will be x% of the times inside this interval. The percentage is generally 95% but it can be 90% or 99%. With the confidence, we specify the width of the interval.

Estimation of the sample size to determine the prevalence and its confidence interval The formulas with the blue background are from a Super Course chapter by R. Heberto Ghezzo, McGill University.

We want to estimate the prevalence of asthma in city XXX. Since diagnosing asthma implies some tests and questionnaires, it is expensive. So we really want to use as few subjects as possible. First question: What precision do we want in our estimate? Let’s say plus/minus 2%. Second question: What is our guess of the prevalence? If we do not know it, better to err on the safe side and assume 50%. Example: required sample size to estimate a prevalence

Worst case scenario: p ~ 50%; 95% CI [48% - 52 %] n = z 2 1-  /2 * p * ( 1.- p)/( d 2 ) n = 1.96 * 1.96 * 0.5 * ( ) / ( 0.04 * 0.04) n = 600 From historical data or another similar city we guess the prevalence: From historical data or another similar city we guess the prevalence: p ~ 0.10; 95% CI [8% - 12 %], then n = 1.96 * 1.96 * 0.1* ( ) / ( 0.04 * 0.04) n = 216 Some information is needed to estimate the sample size. The worst case scenario gives a sample size that is too large.

If the objective is a comparison of two proportions, then we have to specify the level of errors that we can tolerate and the size in absolute units of the smallest difference worth detecting. This is the smallest value that can be credibly attributed to an effect and not to noise, or just random variation. For example, a difference of 5% in the prevalence of a ‘rare’ disease, or an Odds Ratio of, for instance, OR=1.5. Sample Size for Surveys - Comparisons

Example: required sample sizes to compare 2 prevalences (p1 and p2)

We want to test whether the prevalences of a disease is identical in two population with a different exposure. First we must guess which difference we intend to detect. The prevalence in the unexposed group is estimated at 10%, in the exposed group 15%. Second we need to define error rates, alpha : we can easily choose 0.05 as everybody else. beta : here we have a problem. If we are fairly sure that the prevalences will be different, then beta has almost no role in the testing procedure and we can choose a standard value of Example: Sample sizes when testing 2 prevalences

With a power of 80% (beta =.20) we calculate: n = {1.96 * SQRT(.125*.875) *SQRT(.10* *.85)} 2 /{.05} 2 n = 686 (program SSize: n=686) If no difference is a real possibility, then we need a high power (e.g. beta= 0.05). N = 1134 (program SSize: n=1135) We need a substantially larger sample size in each group to be able to say that the difference is less than expected with only a 5 % margin of error.

Required sample sizes when estimating an odds ratio

Example: Sample sizes when testing an odds ratio (1) Example: Sample sizes when testing an odds ratio (1) (What kind of study is this?) Is living downwind from a factory a risk factor for, say, allergic rhinitis? The general population, i.e. upwind from the factory has an exposure proportion of around 4% due to shifting wind directions etc; and a prevalence of rhinitis estimated at 15%. The downwind section has a prevalence of rhinitis estimated at 30% and an exposure of 15%. We do not know the real values—these are guesses. We are interested in an OR of 2.0 with a 95% CI from

Example: Sample sizes when testing an odds ratio (2) We do not know the proportion of exposure in the cases. Hence, first we need to compute p1 (proportion of exposure in the cases). From the formula we get: p1 = p2 * [ OR - p2 * (OR - 1)] We can calculate p2, which is the proportion of exposure in non-cases (e.g. the total population of both regions). If we sample equally from up- and down-wind sections, then we receive the same number of cases and controls from clinics in both sections of the town, the prevalence of exposure in non cases - i.e. controls - can be estimated by:

Example: Sample sizes when testing an odds ratio (3) p2 = (prop. exposed * the prop. diseased in the upwind region + prop. exposed * the prop. diseased in the downwind region) (prop. diseased upwind + prop. diseased downwind) p2= [.04 * *.70 ] / [ ] = Now we calculate p1: p1 = p2 * [ OR - p2 * (OR - 1)] Based on our assumption of an OR=2.0, the estimated proportion of exposed cases is: p1 = * [ * ( 2 - 1)] p1 = 0.17

Example: Sample sizes when testing an odds ratio (4) Now we compute the required case sample size. n = z 2 1-a/2 *{1/[p1*(1-p1)] + 1/[p2*(1-p2)]/ ln 2 (1-e) n = 1.96 * 1.96 * [1/( )+1/(0.089*0.911)]/(ln 2 (0.5) n = 162 Therefore we need 162 subjects with rhinitis and 162 without. To get the 162 with disease we would need, say,  500 subjects from downwind ( 500 * 0.3 = 150 cases) and  500 from upwind ( 500 * 0.04 = 20 cases) to obtain the 170 cases. There will be more than enough controls.

Example: Sample sizes when testing an odds ratio (5) Sample sizes to estimate a Relative Risk using an OR The Relative Risk can be estimated from follow-up studies. Here p1 and p2 refer to incidences, not prevalences. We need some estimate of the annual rate of disease in exposed and non exposed and to proceed as per Odds Ratio. Use PS program. Testing that the Odds Ratio is larger than 1. Testing that the OR > 1 versus OR = 1 is the same as testing p1 > p2 vs p1 = p2. So everything about testing 2 prevalences applies here.

Sample Size for Surveys - true size (1) The formulae presented are theoretical and are based on two assumptions: A. A.A uniform distribution of the variable in the population (no confounders), and B. B.A knowledge of the variability of the variable in the population. These two assumptions are never true. Therefore the calculated sample size is subject to error. We must take this error into account in interpreting the result of the formulae. C. We also need to consider the proportion of participation and attrition.

  As it is better to err by excess, always choose the largest value between the estimators of the population variance, and use a beta error smaller than really needed.   Thus if you err at least the estimated will have the precision required or more.   Cost will be higher, but better to spend a little more in observing more subjects than have a set of estimates that are valueless because of their large confidence intervals. Sample Size for Surveys - true size (2)

Demonstration of SSize.exe 1. 1.Double click on Ssize.exe 2. 2.Choose the scenario that you are interested in. a) Check one example to understand the scenario. b) Evaluate that example. You will see all the formulae. c) Use the table and enter your best guesses. 3.Go directly to( You find it sometimes hidden at the bottom of the page.) Estimate

New approaches can take confounders or multiple risk factors into account Key idea: Central and non-central F- and Chi 2 distribution Assuming the null-hypothesis is true:  = 0 : In other words: We test the Null-Hypothesis. F- and Chi 2 show a central distribution  calculation of the p-value and 95%CI

Assuming the alternative hypothesis is true: > 0 : F- and Chi 2 show a non-central distribution.  calculation of the statistical power

A central and non-central F-distribution (O’Brien, 1986)

Four steps to determine the sample size Four steps to determine the sample size Example: continuous outcome variable (hypertension etc.) Step 1: Outline a scenario Step 2: Estimate the sum of squares (SS) due to the hypothesis (SSH) ` Step 3: Set the  -error Step 4: Calculate the power based on the non-central distribution Least square method: sum of squares about the regression line  F-test = mean SSH / mean square residuals

Step 1: Outline a scenario PB- lower smoking birth proportion of exposure SES weight the target pop. ‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑ ‑ no no no 3300 gram % no no yes 3100 gram % no yes no 3250 gram % no yes yes 3000 gram % yes no no 3250 gram % yes no yes 3000 gram % Outline a scenario for the planned study. Example: potential effects of maternal lead (Pb) exposure on birth weight:

Step 2: Estimate the sums of squares (SS) due to the hypothesis (SSH) Step 3: Set the  -error and a combined standard deviation

Step 4: Calculate the power based on the non- central distribution Reduction of the birth weight of 50 gram sample Lambda Statistical size SSH(population) DFerror Fcritical (FNC) Power N = , N = , N = , N = , N = , N = , N = , N = ,

Summary (SAS power estimation) If we can outline a scenario of potential distributions of risk factors, outcomes, and confounders, then we can use analytical statistical models for the estimation of the sample size. The advantage is that we can already control for potential distortion due to other risk factors that are not included in traditional power estimations.

Demonstration of PS.exe for time-to-event outcomes 1. 1.Double click on PS.exe 2. 2.Click “Overview” or “Continue” 3. 3.Choose the scenario that you are interested in (e.g. Survival). a) Follow the flow of questions (see script for homework 1) b) Calculate the sample size and graph your results. The sample size is for which group? c) Change some assumptions and check the changes in the sample size in the graph.

Sample Size for Surveys - true size (3) Non-Response Non-response is a nuisance, not only because the missing responses but because the sample size gets smaller. If you surveyed 500 subjects and 400 responded, the standard error of the percentage is increased by 12 % because of the non response.

Inputation If the non-response is small, some people suggest to impute a number for the missing responses. (cave !) This 'fabrication of data' is permissible if the only reason for it is the performance of some multivariate technique which requires complete sample. (I doubt it.) Obviously it cannot be used to reduce the error of the estimators. What are limitations for the ‘fabrication of data’??? Sample Size for Surveys - true size (4) Non-Response

Some planning is better than no planning. Remember that the estimated sample sizes are just approximations (best guestimations). The values used in the formulae given are estimations and subject to error. The new sample will rarely have the same prevalence as the other samples from where the assumptions were taken. Always try to err in a conservative way, i.e. estimate a sample size greater than what is really needed. Excess precision is not bad, only expensive. What if the number of cases is restricted? Low precision is not only bad but a waste of resources, work, and time. And the results obtained have little use for another application.