AADAPT Workshop South Asia Goa, December 17-21, 2009 Nandini Krishnan 1.

Slides:



Advertisements
Similar presentations
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation Muna Meky Impact Evaluation Cluster, AFTRL Slides by Paul J.
Advertisements

The World Bank Human Development Network Spanish Impact Evaluation Fund.
Designing an impact evaluation: Randomization, statistical power, and some more fun…
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
AADAPT Workshop South Asia Goa, December 17-21, 2009 Kristen Himelein 1.
Choosing the level of randomization
Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
How to randomize? povertyactionlab.org. Quick Review: Why Randomize Choosing the Sample Frame Choosing the Unit of Randomization Options How to Choose.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Benefits and limits of randomization 2.4. Tailoring the evaluation to the question Advantage: answer the specific question well – We design our evaluation.
The counterfactual logic for public policy evaluation Alberto Martini hard at first, natural later 1.
What could go wrong? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop Africa Program for Education.
1 Managing Threats to Randomization. Threat (1): Spillovers If people in the control group get treated, randomization is no more perfect Choose the appropriate.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
EXPERIMENTS AND OBSERVATIONAL STUDIES Chance Hofmann and Nick Quigley
SAMPLING AND STATISTICAL POWER Erich Battistin Kinnon Scott Erich Battistin Kinnon Scott University of Padua DECRG, World Bank University of Padua DECRG,
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
1 Randomization in Practice. Unit of randomization Randomizing at the individual level Randomizing at the group level –School –Community / village –Health.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Nandini Krishnan Africa Impact Evaluation Initiative World Bank April 14, 2009.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Measuring Impact: Experiments
How do we Randomize? Jenny C. Aker Tufts University.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
Why Use Randomized Evaluation? Presentation by Shawn Cole, Harvard Business School and J-PAL Presenter: Felipe Barrera-Osorio, World Bank 1 APEIE Workshop.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
Impact Evaluation Designs for Male Circumcision Sandi McCoy University of California, Berkeley Male Circumcision Evaluation Workshop and Operations Meeting.
CHAPTER 9: Producing Data: Experiments. Chapter 9 Concepts 2  Observation vs. Experiment  Subjects, Factors, Treatments  How to Experiment Badly 
Why Use Randomized Evaluation? Isabel Beltran, World Bank.
Applying impact evaluation tools A hypothetical fertilizer project.
Impact Evaluation for Evidence-Based Policy Making
What is randomization and how does it solve the causality problem? 2.3.
Measuring Impact 1 Non-experimental methods 2 Experiments
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
TRANSLATING RESEARCH INTO ACTION What and How to randomize? July 9, 2011 Dhaka Raymond Guiteras, Assistant Professor University of Maryland povertyactionlab.org.
CHOOSING THE LEVEL OF RANDOMIZATION. Unit of Randomization: Individual?
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Laura Chioda.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 Steps in Implementing an Impact.
AADAPT Workshop South Asia Goa, India, December 17-21, 2009 Arianna Legovini Manager, Development Impact Evaluation Initiative The World Bank.
Implementing an impact evaluation under constraints Emanuela Galasso (DECRG) Prem Learning Week May 2 nd, 2006.
Randomized Assignment Difference-in-Differences
Impact Evaluation Using Impact Evaluation for Results Based Policy Making Arianna Legovini Impact Evaluation Cluster, AFTRL Slides by Paul J. Gertler &
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Social Experimentation & Randomized Evaluations Hélène Giacobino Director J-PAL Europe DG EMPLOI, Brussells,Nov 2011 World Bank Bratislawa December 2011.
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Using Randomized Evaluations to Improve.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Common Pitfalls in Randomized Evaluations Jenny C. Aker Tufts University.
Development Impact Evaluation Initiative Innovations in investment climate reforms Paris, Nov 13, 2012 In collaboration with the Investment Climate Global.
IMPACT EVALUATION PBAF 526 Class 5, October 31, 2011.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Why Use Randomized Evaluation?
Using Randomized Evaluations to Improve Policy
Measuring Results and Impact Evaluation: From Promises into Evidence
An introduction to Impact Evaluation
Chapter 13- Experiments and Observational Studies
Quasi-Experimental Methods
Using Randomized Evaluations to Improve Policy
Development Impact Evaluation in Finance and Private Sector
Impact Evaluation Toolbox
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Impact Evaluation Designs for Male Circumcision
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Steps in Implementing an Impact Evaluation
Presentation transcript:

AADAPT Workshop South Asia Goa, December 17-21, 2009 Nandini Krishnan 1

 To Identify the causal effect of an intervention  separate the impact of the program from other factors  Need to find out what would have happened without the program  Cannot observe the same person with and without the program at the same point of time >> Create a valid counterfactual 2

High Yield Fertilizer Use OR ? 2) ? 1) ? High Yield Knowledge of Technologies Fertilizer Use

4 (+6) BIASED measure of the program impact

 Hard to distinguish causation from correlation from statistical analysis of existing data  However complex the statistics can only see that X moves with Y  Hard to correct for unobserved characteristics, like motivation/ability  May be very important- also affect outcomes of interest  Selection bias a major issue for impact evaluation  Projects started at specific times and places for particular reasons  Participants may be selected or self-select into programs  First farmers to adopt a new technology are likely to be very different from the average farmer, looking at their yields will give you a misleading impression of the benefits of a new technology 5

6 (+4) Impact of the program (+2) Impact of other (external) factors

 All those in the study have the same chance of being in the treatment or comparison group  By design, treatment and comparison have the same characteristics (observed and unobserved), on average  Only difference is treatment  With large sample, all characteristics average out  Unbiased impact estimates 7

 Lottery (0nly some receive)  Lottery to receive information about a new agricultural technology  Random phase-in (everyone gets it eventually)  Train some farmers groups each year  Variation in treatment  Some get information on new seed, others get credit, others get demonstration plot on their land etc  Encouragement design  One farmers support center per district  Some farmers get travel voucher to attend the center 8

Must receive the program Not suitable for the program Randomize who gets the program

 Budget constraint prevents full coverage  Random assignment (lottery) is fair and transparent  Limited implementation capacity  Phase-in gives all the same chance to go first  No evidence on which alternative is best  Random assignment to alternatives with equal ex ante chance of success 10

 Take up of existing program is not complete  Provide information or incentive for some to sign up- Randomize encouragement  Pilot a new program  Good opportunity to test design before scaling up  Operational changes to ongoing programs  Good opportunity to test changes before scaling them up 11

 Individual  Farm  Farmers’ Association  Irrigation block 12  Village level  Women’s association  Youth groups  School level

 If a program impacts a whole group-- usually randomize whole community to treatment or comparison  Easier to get big enough sample if randomize individuals Individual randomizationGroup randomization

 Randomizing at higher level sometimes necessary:  Political constraints on differential treatment within community  Practical constraints—confusing to implement different versions  Spillover effects may require higher level randomization  Randomizing at group level requires many groups because of within community correlation 14

15 Random assignment Treatment GroupControl Group Participants  Non-participants Evaluation sample Potential participants Maize producersRice producers Target population Commercial Farmers

 External validity  The evaluation sample is representative of the total population  The results in the sample represent the results in the population  We can apply the lessons to the whole population  Internal validity  The intervention and comparison groups are truly comparable   estimated effect of the intervention/program on the evaluated population reflects the real impact on that population 16

 An evaluation can have internal validity without external validity  Example: A randomized evaluation of encouraging women to stand for elections in urban areas may not tell us much about impact of a similar program in rural areas  An evaluation without internal validity, can’t have external validity  If you don’t know whether a program works in one place, then you have learnt nothing about whether it works elsewhere. 17

18 Random Sample- Randomization Randomization National Population Samples National Population

19 Stratification Randomization Population Population stratum Samples of Population Stratum

20 National Population Biased assignment USELESS! Randomization

 Efficacy  Proof of concept  Smaller scale  Pilot in ideal conditions  Effectiveness  At scale  Prevailing implementation arrangements -- “real life”  Higher or lower impact?  Higher or lower costs? 21

 Clear and precise causal impact  Relative to other methods  Much easier to analyze- Difference in averages  Cheaper (smaller sample sizes)  Easier to explain  More convincing to policymakers  Methodologically uncontroversial 22

Wheat plants do NOT  Raise ethical or practical concerns about randomization  Fail to comply with Treatment  Find a better Treatment  Move away—so lost to measurement  Refuse to answer questionnaires  Human beings can be a little more challenging!

 Some interventions can’t be assigned randomly  Partial take up or demand-driven interventions: Randomly promote the program to some  Participants make their own choices about adoption  Perhaps there is contamination- for instance, if some in the control group take-up treatment 24

 Those who get promotion are more likely to enroll  But who got promotion was determined randomly, so not correlated with other observables/non-observables  Compare average outcomes of two groups: promoted/not promoted  Effect of offering the encouragement (Intent-To- Treat)  Effect of the intervention on the complier population (Local Average Treatment Effect) ▪ LATE= ITT/proportion of those who took it up

Assigned to treatment Assigned to control DifferenceImpact: Average treatment effect on the treated Non-treated Treated Proportion treated 100%0%100% Impact of assignment 100% Mean outcome Intent-to-treat estimate 23/100%=23 Average treatment on the treated

Randomly Encouraged Not encouragedDifferenceImpact: Average treatment effect on compliers Non-treated (did not take up program) Treated (did take up program) Proportion treated 70%30%40% Impact of encouragement 100% Outcome Intent-to-treat estimate 8/40%=20 Local average treatment effect

 Calculating sample size incorrectly  Randomizing one district to treatment and one district to control and calculating sample size on number of people you interview  Collecting data in treatment and control differently  Counting those assigned to treatment who do not take up program as control—don’t undo your randomization!! 28

 The treatment already assigned and announced and no possibility for expansion of treatment  The program is over (retrospective)  Universal take up already  Program is national and non excludable  Freedom of the press, exchange rate policy (sometimes some components can be randomized)  Sample size is too small to make it worth it 29

30