Sample size calculation Ahmed Hassouna, MD

Slides:



Advertisements
Similar presentations
The equivalence trial Didier Concordet NATIONAL VETERINARY S C H O O L T O U L O U S E.
Advertisements

LSU-HSC School of Public Health Biostatistics 1 Statistical Core Didactic Introduction to Biostatistics Donald E. Mercante, PhD.
Issues About Statistical Inference Dr R.M. Pandey Additional Professor Department of Biostatistics All-India Institute of Medical Sciences New Delhi.
Confidence Intervals © Scott Evans, Ph.D..
Chapter 11: Sequential Clinical Trials Descriptive Exploratory Experimental Describe Find Cause Populations Relationships and Effect Sequential Clinical.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 9: Hypothesis Tests for Means: One Sample.
Today Concepts underlying inferential statistics
Sample Size Determination
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Basic Statistics in Clinical Research Slides created from article by Augustine Onyeaghala (MSc, PhD, PGDQA, PGDCR, MSQA,
 Mean: true average  Median: middle number once ranked  Mode: most repetitive  Range : difference between largest and smallest.
T tests comparing two means t tests comparing two means.
Statistical Power and Sample Size Calculations Drug Development Statistics & Data Management July 2014 Cathryn Lewis Professor of Genetic Epidemiology.
EDUC 200C Friday, October 26, Goals for today Homework Midterm exam Null Hypothesis Sampling distributions Hypothesis testing Mid-quarter evaluations.
Introduction To Biological Research. Step-by-step analysis of biological data The statistical analysis of a biological experiment may be broken down into.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT OSMAN BIN SAIF Session 19.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
통계적 추론 (Statistical Inference) 삼성생명과학연구소 통계지원팀 김선우 1.
Medical Statistics as a science
Fall 2002Biostat Statistical Inference - Proportions One sample Confidence intervals Hypothesis tests Two Sample Confidence intervals Hypothesis.
Overview and Common Pitfalls in Statistics and How to Avoid Them
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Hypothesis Testing Introduction to Statistics Chapter 8 Feb 24-26, 2009 Classes #12-13.
1 Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
Methods of Presenting and Interpreting Information Class 9.
HYPOTHESIS TESTING.
Chapter 13 Logistic regression.
Statistics for Managers Using Microsoft® Excel 5th Edition
Sample Size Considerations
Sample Size Determination
Biostatistics Case Studies 2007
Section Testing a Proportion
Hypothesis Testing.
Hypothesis and research questions
The Importance of Adequately Powered Studies
Chapter 25: Paired t-Test
Dr. Amjad El-Shanti MD, PMH,Dr PH University of Palestine 2016
Statistical Core Didactic
Statistics for the Social Sciences
Warm Up Check your understanding P. 586 (You have 5 minutes to complete) I WILL be collecting these.
Confidence Intervals and p-values
HYPOTHESIS TESTING Asst Prof Dr. Ahmed Sameer Alnuaimi.
CHAPTER 6 Statistical Inference & Hypothesis Testing
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Chapter 12 Tests with Qualitative Data
Sample Size Estimation
Statistics.
Chapter 8: Inference for Proportions
Central Limit Theorem, z-tests, & t-tests
Crucial Statistical Caveats for Percutaneous Valve Trials
Inferential Statistics
Chapter 11 Goodness-of-Fit and Contingency Tables
One-Way Analysis of Variance: Comparing Several Means
Hypothesis testing. Parametric tests
Review: What influences confidence intervals?
Comparing Populations
The Scientific Method.
Elements of a statistical test Statistical null hypotheses
Introduction to Hypothesis Testing
Sampling and Power Slides by Jishnu Das.
Inferential Statistics
If the chance of having a boy or girl were the same, would you expect there to be more days on which at least 60% of the babies born were boys in a large.
Statistical significance using p-value
Psych 231: Research Methods in Psychology
Chapter Fifteen Frequency Distribution, Cross-Tabulation, and
Chapter 9 Hypothesis Testing.
Type I and Type II Errors
STA 291 Spring 2008 Lecture 17 Dustin Lueker.
Introduction To Hypothesis Testing
Presentation transcript:

Sample size calculation Ahmed Hassouna, MD Sample size calculation Ahmed Hassouna, MD. Diploma of biostatistics applied on clinical trials CESAM, Paris 6 University, France.

Risks of error in statistics: concluding upon a random event is a double jeopardy:

Group 1 TTTa Group 2 TTTb We begin by assuming that study groups are comparable BEFORE the beginning of the study. Hence, any difference collected AFTER study, can be attributed to effect treatment.

1- We can conclude upon a difference that does not exist. TTTa Group 2 TTTb The risk to conclude upon an evidence that “does not exist”. The difference recorded AFTER the study was already present BEFORE the study; i.e. groups were not initially (primarily) non comparable from the start. Primary risk of error (α)

2- We can miss an existing difference General hospital Small hospital If the known percentage of baby boys to baby girls is 1:1, in which hospital do you think that one day will be born as much as twice baby girls to baby boys? Skewed data are more probable in a small sample than in large ones; carrying the risk “to miss” an existing relation. Secondary risk of error (β) 100 boys/ 100 girls 1-3 boys and girls Absence of evidence is not an evidence of absence

Factors governing sample size calculation 1- How much likely we want to be wrong in our conclusion (α) ? 2- How much likely we are ready to miss a conclusion (β) ? 3- What is the magnitude of the difference that we want to put into evidence? 4- What is the variability of outcome measure ? 5- What is the strength of our comparator: a competitor or just a placebo ?

1- How likely we want to make a “wrong” conclusion? What is the limit of our P value? We have to accept the “universal limit of being wrong in 5% of cases. Why? Because this is the maximum limit allowed to give to the world “a useless or harmful drug”

2- How likely we want to miss the evidence? The power of a study is less strict: 80-90%...Why? Because it is a question of money. Because the produced harm is less than that associated with the primary risk of error; i.e. missing a “potentially beneficial drug”. Reality (+) (-) Study (+) α (-) β Power of the study

3- Magnitude of the difference “The classic example” is the comparison of “THE CURE” and “THE POISON”. How many patients are you willing to sacrifice to show the huge difference between giving and taking life? “The real life” is that the differences are usually small and, the smaller is the difference; the larger your sample size should be. Always remember the mega trials arranged by pharmaceutical companies.

4- Variability of outcome The extreme example is that any patient taking the treatment will survive and every one who will be deprived from it will eventually die. The common example is that some patients improve on treatment, others show “some” improvement and “few” patients do not improve at all. The more is the variability of outcome, the more patients we will have to include in our study.

5- The direction of the study A bilateral study compares 2 treatments; aiming to show which is better: A or B. A unilateral study is usually organized to test the efficiency of a treatment by comparing it to placebo. It offers the opportunity to reach statistical significance with a smaller difference that is achieved with a smaller sample.

In order to maintain a total α of 5% α/2 = 2.5% α/2 = 2.5% α = 5% The IC at 95% The IC at 95% m m We are interested to know whether the birth weight is significantly different (smaller or larger) from the rest of the group; which is the typical question posed by a bilateral design: is “A” significantly better than “B” or significantly worse? In a unilateral design, our only interest is to know if the child is significantly heavier than the other children? A typical treatment vs. placebo study is only interested to prove that treatment is better than placebo but never the reverse.

In order to maintain a total α of 5% α/2 = 2.5% α/2 = 2.5% α = 5% The null hypothesis The null hypothesis 2 SD 1.65 SD m m H0 = both treatments are comparable. H1= TTTa is better than TTTb or TTTb is better than TTTa. The critical point of rejection is 2 SD from either sides of mean. H0= both treatments are comparable + placebo is better H1= treatment is better than placebo The critical point of rejection is 1.65 SD on one side only.

Single Outcome measure HF/not Single Outcome measure Qualitative Ordinal Categorical Number of groups 1 group 2 groups Multiple groups Number of measurements Once Repeated Quantitative Time-related event NYHA class peptide CRT

How to intelligently reduce your sample size? d/SD

Maximize the target difference (d) Avoid choosing CRUDE OUTCOMES as an end point. As example taking mortality as an indicator for successful myocardial preservation during CABG or PCI will be a total failure; for the fact that mortality figures are as low as 1-2% with only decimals differences, regardless of the preservation techniques or type of stent. A more sensitive measure of ischemia as Troponin I or a score made of a large number of indicators of morbidity (Syntax score) is capable to show larger detectable differences among compared groups.

Reduce variability (SD) Randomization and Stratification that allows the even distribution of sources of variability between compared groups. Adopting a unilateral or cross-over design if feasible or by reducing power of the study from 80 to 70%; which reduces size by 20%? Blocking some sources of variation and narrowing the scope of the study, like excluding females, known for bad outcome after CABG, from a minimal resources study on the subject.

Reduce variability (SD) Choosing a low variability outcome. Transforming a binary outcome into an ordered categorical one, reduces sample size by 30%. Repeatedly measuring a quantitative outcome by comparing pre to post values, taking the average of post values or better using the repeated measures ANOVA procedure.

Online sample size calculators. Survey study: http://www.raosoft.com/samplesize.html Comparison of 2 means or 2 proportions: http://www.stat.ubc.ca/~rollin/stats/ssize/n2.html http://www.stat.ubc.ca/~rollin/stats/ssize/b2.html Relative risk: http://www.stat.ubc.ca/~rollin/stats/ssize/caco.html.

Online sample size calculators Correlation or regression: www.hedwig.mgh.harvard.edu/sample_size/js/js_associative_quant.html Cross-over study: http://hedwig.mgh.harvard.edu/sample_size/js/js_crossover_quant.html

Thank you