Sample Size and Statistical Power Epidemiology 655 Winter 1999 Jennifer Beebe.

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
Hypothesis Testing making decisions using sample data.
Chapter 10.  Real life problems are usually different than just estimation of population statistics.  We try on the basis of experimental evidence Whether.
Statistical Issues in Research Planning and Evaluation
ProportionMisc.Grab BagRatiosIntro.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Cal State Northridge  320 Ainsworth Sampling Distributions and Hypothesis Testing.
Research Curriculum Session III – Estimating Sample Size and Power Jim Quinn MD MS Research Director, Division of Emergency Medicine Stanford University.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
Chapter 9 Hypothesis Testing.
BCOR 1020 Business Statistics
Today Concepts underlying inferential statistics
Using Statistics in Research Psych 231: Research Methods in Psychology.
Sample Size Determination
Inferential Statistics
Hypothesis Testing. Outline The Null Hypothesis The Null Hypothesis Type I and Type II Error Type I and Type II Error Using Statistics to test the Null.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Tests of Hypotheses Based on a Single Sample.
Inferential Statistics
Thomas Songer, PhD with acknowledgment to several slides provided by M Rahbar and Moataza Mahmoud Abdel Wahab Introduction to Research Methods In the Internet.
Sample size calculation
AM Recitation 2/10/11.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Introduction to Biostatistics and Bioinformatics
Tuesday, September 10, 2013 Introduction to hypothesis testing.
Tests of significance & hypothesis testing Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics.
Hypothesis Testing (Statistical Significance). Hypothesis Testing Goal: Make statement(s) regarding unknown population parameter values based on sample.
Hypothesis Testing.
Multiple Choice Questions for discussion
OKU 9 Chapter 15: ORTHOPAEDIC RESEARCH Brian E. Walczak.
1 Today Null and alternative hypotheses 1- and 2-tailed tests Regions of rejection Sampling distributions The Central Limit Theorem Standard errors z-tests.
Jan 17,  Hypothesis, Null hypothesis Research question Null is the hypothesis of “no relationship”  Normal Distribution Bell curve Standard normal.
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
LECTURE 19 THURSDAY, 14 April STA 291 Spring
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Hypotheses tests for means
Lecture 16 Dustin Lueker.  Charlie claims that the average commute of his coworkers is 15 miles. Stu believes it is greater than that so he decides to.
No criminal on the run The concept of test of significance FETP India.
Statistics (cont.) Psych 231: Research Methods in Psychology.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
The use & abuse of tests Statistical significance ≠ practical significance Significance ≠ proof of effect (confounds) Lack of significance ≠ lack of effect.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 8 Hypothesis Testing.
Chapter 20 Testing Hypothesis about proportions
Lecture 18 Dustin Lueker.  A way of statistically testing a hypothesis by comparing the data to values predicted by the hypothesis ◦ Data that fall far.
Education 793 Class Notes Decisions, Error and Power Presentation 8.
Using Inference to MAKE DECISIONS The Type I and Type II Errors in Hypothesis Testing.
Lecture 17 Dustin Lueker.  A way of statistically testing a hypothesis by comparing the data to values predicted by the hypothesis ◦ Data that fall far.
Module 15: Hypothesis Testing This modules discusses the concepts of hypothesis testing, including α-level, p-values, and statistical power. Reviewed.
Math 4030 – 9a Introduction to Hypothesis Testing
© Copyright McGraw-Hill 2004
URBDP 591 I Lecture 4: Research Question Objectives How do we define a research question? What is a testable hypothesis? How do we test an hypothesis?
1.  What inferential statistics does best is allow decisions to be made about populations based on the information about samples.  One of the most useful.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Hypothesis Testing Steps for the Rejection Region Method State H 1 and State H 0 State the Test Statistic and its sampling distribution (normal or t) Determine.
Chapter 9: Hypothesis Tests for One Population Mean 9.2 Terms, Errors, and Hypotheses.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Inferential Statistics Psych 231: Research Methods in Psychology.
Chapter 9 Hypothesis Testing Understanding Basic Statistics Fifth Edition By Brase and Brase Prepared by Jon Booze.
Learning Objectives Describe the hypothesis testing process Distinguish the types of hypotheses Explain hypothesis testing errors Solve hypothesis testing.
Sample size calculation
Hypothesis Testing I The One-sample Case
Understanding Results
Sample Size Estimation
Testing Hypotheses I Lesson 9.
AP STATISTICS LESSON 10 – 4 (DAY 2)
Rest of lecture 4 (Chapter 5: pg ) Statistical Inferences
Presentation transcript:

Sample Size and Statistical Power Epidemiology 655 Winter 1999 Jennifer Beebe

Determining Sufficient Sample Size Purpose: To provide an understanding of the concepts of sample size and statistical power; to provide tools for sample size calculation

Why do we worry about Sample Size and Power? Sample size too big; too much power wastes money and resources on extra subjects without improving statistical results Sample size too small; having too little power to detect meaningful differences –exposure (treatment) discarded as not important when in fact it is useful Improving your research design Improving chances for funding

Review of Statistical Concepts Hypothesis testing –Null hypothesis H o : No difference between groups; no effect of the covariate on the outcome –Alternative hypothesis H a : The researcher’s theory –Decision rule: Reject H o if a test statistic is in the critical region (p<.05)

Hypothesis Testing: Example H o : Diabetes is not associated with endometrial cancer in postmenopausal women H a : –Diabetes is associated with endometrial cancer; direction of association not specified (two-sided test) –Women with diabetes have an increased risk of developing endometrial cancer (one-sided test) –Women with diabetes have a decreased risk of developing endometrial cancer (one-sided test)

Under optimal conditions, we would examine all postmenopausal women with and without diabetes to determine if diabetes is associated with endometrial cancer –Instead, we collect data on a sample of postmenopausal women –Based on sample data, we would conduct a statistical test to determine whether or not to reject the null hypothesis

Errors Our sample may not accurately reflect the target population and we may draw an incorrect conclusion about all postmenopausal women based on the data obtained from our sample Type I and Type II errors

Two Types of Error Type I: Rejecting the H o when H o is true –The probability of a Type I error is called  –  is the designated significance level of the test –Usually we set the critical value so  =0.05 In our example, we could conclude based on our sample, that diabetes is associated with endometrial cancer when there really is no association

P-values Measure of a Type I error (random error) Probability that you have obtained your study results by chance alone, given that your null hypothesis is true If p=0.05, there is just a 5% chance that an observed association in your sample is due to random error

Example:Diabetes and Endometrial Cancer From our sample data, we found that women who have diabetes are 2 times more likely to develop endometrial cancer when compared to women without diabetes (p=0.01) If diabetes and endometrial cancer are not associated, there is a 1% probability that we would find this association by chance if we set the critical value as 0.05; 0.01<0.05; we would reject H o in favor of H a

Type II Error Type II: Accept H o when H a is true The probability of a type II error is called   depends on the effect size (How far from H o are we?) If we are far from H o, then  is small If we are close to H o, then  is large In our example, we could conclude that there is no association between diabetes and endometrial cancer when in fact there is an association

Truth in the Population AssociationNo association Studyb/w predictorb/w predictor Resultsand outcomeand outcome Reject HoCorrectType I error Fail to Type II errorCorrect Reject Ho

Power Power is the probability of observing an effect of a particular magnitude in the sample if one of a specified effect size or greater actually exists in the population Power = 1-  if  =.20 then power =.80; we will accept a 20% chance of missing an association of a particular size b/w an exposure and an outcome if one really exists

 and  Levels Usually range from (  ) and from (  ) Convention  =0.05 and  =0.20 Use low alpha’s to avoid false positives Use low beta’s to avoid false negatives Increased sample size will reduce type I and type II errors

Asking the sample size question? What sample size do I need to have adequate power to detect a particular effect size (or difference)? I only have N subjects available. What power will I have to detect a particular effect size (or difference) with that sample size?

Preparing to Calculate Sample Size What kind of study are you doing? –Case-control, cross-sectional, cohort What is the main purpose of the study? –What question(s) are you asking? What is your outcome measure? –Is it continuous, dichotomous, ordinal? The prevalence of exposure(s) in study population?

Preparing to Calculate Sample Size What statistical tests will be used? –(t-test, ANOVA, chi-square, regression etc) Will the test be one or two tailed? What  level will you use? –  =0.05 The hard one: How small an effect size (or difference) is important to detect? –What difference would you not want to miss? With what degree of certainty (power) do you want to detect the effect? (80-95%)

Tradeoffs with Sample Size Sample size is affected by effect size, , , power If detected effect size is  (Big OR or RR) then sample size  If detected effect size is  (Small OR or RR) then sample size  If the effect size is fixed; –  ;  ; (1-  )  ; sample size 

Tradeoffs with Power Power affected by sample size, prevalence of exposure, , , effect size  sample size;  power  effect size to detect;  power  ;  power Power of study is optimal usually when prevalence of the exposure in the control or referent group is b/w 40-60% Equal numbers of subjects in each group will increase power

Sample Size Requirements in a Cohort / Cross-sectional Study In addition to specified  and power, sample size depends on the – Incidence or probability of outcome among the unexposed –Ratio of exposed / unexposed –Relative risk/prevalence ratio that one regards as important to detect

Sample Size Requirements for a Case-control Study In addition to specified  and power, sample size depends on the –Ratio of cases to controls –Proportion of controls exposed –Odds ratio that one regards as important to detect

Sample Size and Power Software EpiInfo –Programs  Statcalc  Sample size and Power –User-friendly; easily accessible nQuery –More sophisticated, lots of options, you need to supply program with more information PASS, Power and Precision, GPower

Helpful Hints Choose an effect size reasonable for observational studies (this may be based on previous literature) Knowledge of prevalence of exposures of interest (also based on previous literature) Increase sample size 10-20% for each major confounder