The days ahead Monday-Wednesday –Training workshop on how to measure the actual reduction in HIV incidence that is caused by implementation of MC programs.

Slides:



Advertisements
Similar presentations
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Advertisements

Mywish K. Maredia Michigan State University
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
Sample size issues & Trial Quality David Torgerson.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Journal Club Alcohol, Other Drugs, and Health: Current Evidence July–August 2013.
Experimental Design making causal inferences. Causal and Effect The IV precedes the DV in time The IV precedes the DV in time The IV and DV are correlated.
Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.
2.6 The Question of Causation. The goal in many studies is to establish a causal link between a change in the explanatory variable and a change in the.
Public Policy & Evidence: How to discriminate, interpret and communicate scientific research to better inform society. Rachel Glennerster Executive Director.
Non-Experimental designs: Developmental designs & Small-N designs
N The Experimental procedure involves manipulating something called the Explanatory Variable and seeing the effect on something called the Outcome Variable.
The Study of Adult Development and Aging:
PAI786: Urban Policy Class 2: Evaluating Social Programs.
Association vs. Causation
Statistical Issues in Data Collection and Study Design For Community Programs and Research October 11, 2001 Elizabeth Garrett Division of Biostatistics.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
Introduction to Impact Evaluation: What is it and why do we use it? Male Circumcision Impact Evaluation Meeting Johannesburg, South Africa January 18-23,
Research and Evaluation Center Jeffrey A. Butts John Jay College of Criminal Justice City University of New York August 7, 2012 How Researchers Generate.
Impact Evaluation in the Real World One non-experimental design for evaluating behavioral HIV prevention campaigns.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 13 Experiments and Observational Studies.
Measuring Impact: Experiments
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)
 Is there a comparison? ◦ Are the groups really comparable?  Are the differences being reported real? ◦ Are they worth reporting? ◦ How much confidence.
Experimental Design making causal inferences Richard Lambert, Ph.D.
Experimental Design All experiments have independent variables, dependent variables, and experimental units. Independent variable. An independent.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
Impact Evaluation Designs for Male Circumcision Sandi McCoy University of California, Berkeley Male Circumcision Evaluation Workshop and Operations Meeting.
Chapter 3.1.  Observational Study: involves passive data collection (observe, record or measure but don’t interfere)  Experiment: ~Involves active data.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Current practices in impact evaluation Howard White Independent Evaluation Group World Bank.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
1 The Training Benefits Program – A Methodological Exposition To: The Research Coordination Committee By: Jonathan Adam Lind Date: 04/01/16.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
AP Statistics Exam Review Topic #4
CHAPTER 4 Designing Studies
Measuring Results and Impact Evaluation: From Promises into Evidence
Right-sized Evaluation
Fundamentals of Monitoring and Evaluation
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
1 Causal Inference Counterfactuals False Counterfactuals
Impact Evaluation Toolbox
Chapter 4: Designing Studies
CHAPTER 4 Designing Studies
Impact Evaluation Designs for Male Circumcision
CHAPTER 4 Designing Studies
Class 2: Evaluating Social Programs
Class 2: Evaluating Social Programs
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Module appendix - Attributable risk
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Presentation transcript:

The days ahead Monday-Wednesday –Training workshop on how to measure the actual reduction in HIV incidence that is caused by implementation of MC programs Thursday & Friday –Share experiences about the problems countries are encountering in initial scale-up of MC services – and the solutions they have found. Clarify what the unanswered questions are about how best to implement MC. Saturday –Explore how measurement of the impact on incidence could also answer unresolved questions about how to best implement MC (e.g. optimal task sharing…) Monday-Wednesday next week –Third meeting on the use of the DMPPT tool to estimate the cost and forecast the possible impact of MC programs to guide initial program design (and help decision-makers decide whether they want to implement MC)

How can we estimate effectiveness? E.g., What is the effect of: Male circumcision on HIV incidence In other words: “How much does an increase in male circumcision cause HIV incidence to decrease?”

How can we estimate effectiveness? If we implement MC in a community and incidence of HIV does not change does that mean that MC is not effective at reducing incidence of HIV? If we implement MC and incidence goes up, does that mean that MC increases HIV transmission? If we implement MC and incidence falls, do we know that the MC caused the fall?

We are not asking whether incidence changes at the same time that MC is implemented – we are asking whether incidence changes because MC is implemented. Thus, we need to know how much would incidence have changed if MC was not implemented – and how different is that from how much it changes with MC How can we estimate effectiveness?

Unfortunately, we can’t rewind the clock and observe the same community over the same period of time with and without MC We need an estimate what would have happened to HIV incidence in the absence of an MC program We call this the counterfactual –i.e. the “factual” is what really happens where MC is implemented and the “counterfactual” is what would have happened if MC had not been implemented How can we estimate effectiveness?

Creating a Good Counterfactual Sometimes it is very simple: –If I dye my hair tonight and come in tomorrow with black hair you will easily believe that the change in hair color was caused by the hair dye –In other words, if we strongly believe that something is not changing on its own, then we can assume that the counterfactual = baseline –Good example of that for HIV is treatment

Creating a Good Counterfactual With HIV prevention it is never simple

Creating a Good Counterfactual Baseline MC Implemented

Creating a Good Counterfactual Baseline MC Implemented IMPACT

Creating a Good Counterfactual Baseline MC Implemented IMPACT ? MC doubles HIV incidence?

Creating a Good Counterfactual Baseline MC Implemented Real IMPACT MC Halves HIV incidence

Creating a Good Counterfactual Baseline MC Implemented IMPACT? MC reduces HIV incidence by 1/2

Creating a Good Counterfactual Baseline MC Implemented Real IMPACT MC reduces HIV incidence by 1/3

Creating a Good Counterfactual Where in the World Are We?

Creating a Good Counterfactual Where in the World Are We? If the purpose of the experiment was to test the effectiveness of the hints, should we have given the hints to the “experts” or to the rest of us?

Creating a Good Counterfactual Low INCIDENCE High EFFECT Ideally you want the communities that receive MC to be indistinguishable from those that don’t

Creating a Good Counterfactual NEGATIVE EFFECT? If they are different, it can bias the estimation of the effect Low INCIDENCE High

Creating a Good Counterfactual If they are different, it can bias the estimation of the effect HUGE EFFECT? Low INCIDENCE High

Creating a Good Counterfactual The beneficiaries of the intervention and the counterfactual or control groups: –have identical characteristics, except for benefiting from the intervention With a good counterfactual, the only reason for different outcomes between treatments and controls is the intervention (I)

Creating a Good Counterfactual The beneficiaries of the intervention and the counterfactual or control groups: –have identical characteristics, except for benefiting from the intervention With a good counterfactual, the only reason for different outcomes between treatments and controls is the intervention (I)

Creating a Good Counterfactual for MC What if we compared the incidence among the men in a community who decide to get circumcised with the incidence among the men who decide not to get circumcised? What if we compared the incidence in communities who decide to initiate circumcision services in their health center with the incidence in the ones that don’t? What if we compare the incidence in towns without a clinic with the incidence in towns with a circumcision clinic?

Creating a Good Counterfactual for MC What if we opened up a circumcision clinic in Soweto and another in Khayelitsha and another in KwaMashu – but they could only serve a small number of the men who want to be circumcised. If we give lottery tickets to see who can go in what month, could we use the ones who will wait a year as a comparison (counterfactual) group? What if we focused circumcision services on the year olds. Could we use the year olds as a comparison group?

Multiple threats to a valid estimate of effectiveness regardless of a robust counterfactual ( we have to live with these ) Measurement error/biases Endogenous Change –Secular changes or drift (long term trends in community, region or country) –Maturational trends (Individual change) –Interfering events Hawthorne/cohort effects

Multiple threats to a valid estimate of effectiveness that can be controlled for with a robust counterfactual ( examples will be shown examples on slides to come ) Selection Bias –Participants who voluntarily participate or have the exposure may be different than those without Confounding factors that obfuscate attribution of the program to the impact; a robust counterfactual can control for such factors that are known and unknown

Selection bias from an evaluation of “enrolled versus not enrolled”

Designs Leading to Biased Results: “Enrolled versus Not Enrolled” Consider a school-based pregnancy prevention program 10 schools in the district are asked if they would like to participate

Designs Leading to Biased Results: “Enrolled versus Not Enrolled” 5 schools decline participation 5 schools elect to participate in the program Pregnancy Prevention Program No intervention

Designs Leading to Biased Results: “Enrolled versus Not Enrolled” Pregnancy Prevention Program No intervention Pregnancy rate = e t per 100 student years 2 per 100 student years

Pregnancy Prevention Program No intervention Pregnancy rate = 3 per 100 student years 2 per 100 student years Designs Leading to Biased Results: “Enrolled versus Not Enrolled ” Schools in the program had fewer adolescent pregnancies… Can we attribute this difference to the program?

Designs Leading to Biased Results: “Enrolled versus Not Enrolled ” Pregnancy Prevention Program No intervention Pregnancy rate = 3 per 100 student years 2 per 100 student years religiouee erg) Factor X

Designs Leading to Biased Results: “Enrolled versus Not Enrolled” Pregnancy Prevention Program No intervention Pregnancy rate = 3 per 100 student years 2 per 100 student years Factor X (not sure what this should be) Factor X Observed effect might be due to differences in “Factor X” which led to differential selection into the program (“selection bias”)

Designs Leading to Biased Results: “Enrolled versus Not Enrolled ” This design compares “apples to oranges” The reason for not enrolling might be correlated with the outcome –You can statistically “control” for observed factors –But you cannot control for factors that are “unobserved” Estimated impact erroneously mixes the effect of different factors

Confounding that results even from a a “matched design”

: “Matched Designs” Individuals, groups, or communities are matched based on known characteristics to improve comparability: Age, race, sex Region, poverty

“Matched Designs” From each pair, one receives the intervention Differences in outcomes are compared within the pair

: “Matched Designs” Does this design ensure that the matched pairs are comparable on all factors except the intervention? No: Only observed factors are used for matching Unobserved factors may differ

World Bank Challenged: Are Poor Really Helped? By Celia Dugger New York Times, July 28, 2004 WASHINGTON - Wealthy nations and international organizations, including the World Bank, spend more than $55 billion annually to better the lot of the world's 2.7 billion poor people. Yet they have scant evidence that the myriad projects they finance have made any real difference, many economists say.

Not sure where previous slide goes Co we could end here and then play the game that is illustrative of selection bias and then after that, have the over talk on IE study designed and analytic methods. this could be doner as the whole group as opposed to country teams