Estimating the Effects of Treatment on Outcomes with Confidence Sebastian Galiani Washington University in St. Louis.

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 9 Inferences Based on Two Samples.
Advertisements

Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Hypothesis Testing Steps in Hypothesis Testing:
Hypothesis: It is an assumption of population parameter ( mean, proportion, variance) There are two types of hypothesis : 1) Simple hypothesis :A statistical.
Statistical Decision Making
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Independent Samples and Paired Samples t-tests PSY440 June 24, 2008.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
Topic 2: Statistical Concepts and Market Returns
Lecture 9: One Way ANOVA Between Subjects
BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
Independent Sample T-test Often used with experimental designs N subjects are randomly assigned to two groups (Control * Treatment). After treatment, the.
Chapter 2 Simple Comparative Experiments
Today Concepts underlying inferential statistics
The t Tests Independent Samples.
Inferential Statistics
Hypothesis Testing and T-Tests. Hypothesis Tests Related to Differences Copyright © 2009 Pearson Education, Inc. Chapter Tests of Differences One.
Choosing Statistical Procedures
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
AM Recitation 2/10/11.
Chapter 4 Hypothesis Testing, Power, and Control: A Review of the Basics.
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Chapter 8 Introduction to Hypothesis Testing
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
CHAPTER 16: Inference in Practice. Chapter 16 Concepts 2  Conditions for Inference in Practice  Cautions About Confidence Intervals  Cautions About.
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
1 Statistical Inference Greg C Elvers. 2 Why Use Statistical Inference Whenever we collect data, we want our results to be true for the entire population.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Chapter 8 Introduction to Hypothesis Testing
Inferential Statistics 2 Maarten Buis January 11, 2006.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
B AD 6243: Applied Univariate Statistics Hypothesis Testing and the T-test Professor Laku Chidambaram Price College of Business University of Oklahoma.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Introduction to Inferential Statistics Statistical analyses are initially divided into: Descriptive Statistics or Inferential Statistics. Descriptive Statistics.
통계적 추론 (Statistical Inference) 삼성생명과학연구소 통계지원팀 김선우 1.
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005 Dr. John Lipp Copyright © Dr. John Lipp.
5.1 Chapter 5 Inference in the Simple Regression Model In this chapter we study how to construct confidence intervals and how to conduct hypothesis tests.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Statistical Testing of Differences CHAPTER fifteen.
ANOVA Assumptions 1.Normality (sampling distribution of the mean) 2.Homogeneity of Variance 3.Independence of Observations - reason for random assignment.
Copyright ©2013 Pearson Education, Inc. publishing as Prentice Hall 9-1 σ σ.
Education 793 Class Notes Decisions, Error and Power Presentation 8.
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 7 Sampling, Significance Levels, and Hypothesis Testing Three scientific traditions.
Chapter 10 The t Test for Two Independent Samples
T tests comparing two means t tests comparing two means.
© Copyright McGraw-Hill 2004
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
European Patients’ Academy on Therapeutic Innovation The Purpose and Fundamentals of Statistics in Clinical Trials.
Significance Tests for Regression Analysis. A. Testing the Significance of Regression Models The first important significance test is for the regression.
Chapter 13 Understanding research results: statistical inference.
Hypothesis Tests. An Hypothesis is a guess about a situation that can be tested, and the test outcome can be either true or false. –The Null Hypothesis.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Chapter 9 Introduction to the t Statistic
Learning Objectives Describe the hypothesis testing process Distinguish the types of hypotheses Explain hypothesis testing errors Solve hypothesis testing.
Inference about the slope parameter and correlation
Inference and Tests of Hypotheses
Chapter 2 Simple Comparative Experiments
Statistical Power.
Rest of lecture 4 (Chapter 5: pg ) Statistical Inferences
Presentation transcript:

Estimating the Effects of Treatment on Outcomes with Confidence Sebastian Galiani Washington University in St. Louis

Parameters of Interest Two parameters of interest widely used in the literature: Average Treatment Effect Average Treatment Effect on the Treated Under randomization and full compliance, they coincide.

Randomization In the absence of difficulties such as noncompliance or loss to follow up, assumptions play a minor role in randomized experiments, and no role at all in randomized tests of the hypothesis of no treatment effect. In contrast, inference in a nonrandomized experiment requires assumptions that are not at all innocuous.

Quasi-Experimental Designs If randomization is not feasible, we need to rely on quasi-experimental methods. In our case, the most promising strategy would be a Generalized Difference in Differences strategy.

Parameters of Interest We might want to response the following questions: What is the effect of the intervention on a given outcome on a given population? What is the effect of the intervention on a given outcome on those that self-select as users of the facilities? The power for identifying the first parameters might be lower than the power for the second parameter –identified by IV Methods.

Distance to Facilities and Sampling Think about stratifying the sample by distance to facilities, and over-sample households residing near facilities.

Testing absence of Treatment Effects. Type I Error Once we have chosen a Type I error rate: , the null hypothesis (  T -  C = 0) is rejected whenever the statistics of contrast |t| > t*  ; where t*  is the (critical) value of t that defines the  /2 percentile of the distribution of t. 0 -t*  /2 t*  /2 No rejection Reject Null

Type II Error Now consider an alternative hypothesis:  T -  C = d Under this alternative hypothesis, the t-statistic will have a different distribution. If the alternative hypothesis is true, we want to reject the null hypothesis as often as possible. To fail to do so would be a Type II error. We want to restrict the probability of this type of error to  Then  will be the type II error rate of the test. And 1-  will be the power of the statistical test. The power of a statistical hypothesis test measures the test's ability to reject the null hypothesis when it is actually false.

Type II Error 0-t*  /2 t*  /2 No rejection Reject Null Distribution of t under the null Distribution of t under the alternative Power 1-   There is a trade-off between Type I and Type II errors

Type I and II Errors 0-t*  /2 t*  /2 No rejection Reject Null Distribution of t under the null Distribution of t under the alternative Power 1-   Both errors can be simultaneously reduced if the dispersion of the statistics is reduced.

Simple Clinical Trial In this design m members are allocated to each condition: treatment and control. The observed value to the i-th member in the l-th condition is a function of the grand mean and the effect of the l-th condition; any difference between the observed and the predicted value is allocated to the residual error. The intervention effect is C l. Its estimate is:

Simple Clinical Trial Under the null hypothesis, H 0 : C l = 0, Let estimate the variance of. Assuming that the residual error is distributed Gaussian, the intervention effect is evaluated using a t-statistic with the appropriate df. The researcher determines the desired Type I and II error rates (say 5% and 20%, respectively). The researcher expects a negative intervention effect but would be concerned about a positive effect; as a result, she chooses a two-tailed test.

Simple Clinical Trial Given the random assignment of members to treatment and control conditions, it is reasonable to assume that the two study conditions are independent. Then: The estimated variance of a single condition mean is: If we assign the same number of members in each condition, the variances in the two conditions are assumed to be equal.

Simple Clinical Trial Then, the t-statistic is estimated as: The parameters appearing in this formula are relatively easy to estimate using data from previous reports, from analyses of existing data or from preliminary studies.

Simple Clinical Trial True type I and II errors rates will be  and  respectively if: 0-t*  /2 t*  /2 No rejection Reject Null Distribution of t under the null Distribution of t under the alternative Power 1-   -t*  /2 -t*  H0H0 HAHA

Or: This expression is general to any design. We need: –Desired type I and II error rates. –The expected magnitude of the treatment effect. –The expression for the variance of the estimated treatment effect, which is a function of the sample size. We can express any of these variables in terms of the others. General Expression

Simple Clinical Trial Sample size:

Sample Size in GRT Assume that there is only one individual per household. Probability that an individual has diarrhea: Individual i, in group k, assigned to condition l. Within each condition: the variance of any given observation is: Where stands for the variance within groups and for the variance between groups

Sample Size in GRT Consider first the group mean. If that mean were based on m independent observations, the variance of that mean would be estimated as: However, because the members within an identifiable group almost always show positive intra-class correlation, those observations are not independent. In fact, only the variance attributable to the individual effect will vanish as m increases. The variance attributable to the group effect will remain unaffected.

Sample Size Then, the variance of the group mean is: Where, m stands for the number of households per group and ICC for the intra-group correlation. The variance of the condition mean is: Where g is the number of groups in each condition

Sample Size When ICC>0, the variance of the condition mean is always larger in a GRT than in a study based on random assignment of the same number of individuals to the study conditions. Statistic of interest: Variance of the statistic:

Sample Size Given a moderate number of groups per condition, the t-statistic to asses the difference between condition means is: It is distributed t-student with g T +g C -2 degrees of freedom

Sample Size Sample size: –Number of groups per condition: –Number of household per group (it requires a couple of iterations):

Sample Size When each household has more than one observation, we need to perform the following correction: Where a is the number of observations per household and  is the intra-household correlation. See extreme cases  or 

Pretest - Posttest: Repeat Observations on Groups Data are collected in each condition before and after the intervention has been delivered in the intervention condition. There are repeated observations on the same groups

Pretest - Posttest: Repeat Observations on Groups The model: The observed value for the i-th member nested within the k-th group and l-th condition and observed at the j-th time is expressed as a function of the grand mean, the effect of the l-th condition, the effect of the j-th time, the joint effect of condition and time, the realized value of the k- th group, the joint effect of group and time. Differences between this predicted value and the observed value are allocated to the residual error

Pretest- Posttest: Repeat Observations on Groups Treatment effect: Dif-in-dif There are two sources of variation among the groups: –Variation due to group effect –Variation due to the interaction group x time The first difference eliminates the first source of variation. The group mean is: This model can be easily transformed in the basic GRT

Pretest - Posttest: Repeat Observations on Groups The variance of the group mean is: Following the same steps as before… The variance of the intervention effect can be written as: Sample size can be solved as before

Pretest - Posttest: Repeat Observations on Members Data are collected in each condition before and after the intervention has been delivered in the intervention condition. There are repeated observations on the same members

Pretest- Posttest: Repeat Observations on Members The model: The observed value for the i-th member nested within the k-th group and l-th condition and observed at the j-th time is expressed as a function of the grand mean, the effect of the l-th condition, the effect of the j-th time, the joint effect of condition and time, the realized value of the k-th group, the realized value of the i-th member, the joint effect of group and time and the joint effect of member and time. Differences between this predicted value and the observed value are allocated to the residual error

Pretest- Posttest: Repeat Observations on Members Treatment effect: Dif-in-dif There are three sources of variation among the members: –Variation due to member effect –Variation due to the interaction member x time –Error term The first difference eliminates the first source of variation.

Pretest- Posttest: Repeat Observations on Members Taking differences by members: This model can be easily transformed in the basic GRT The variance of the intervention effect can be written as: Sample size can be solved as before