Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Using Randomized Evaluations to Improve.

Slides:



Advertisements
Similar presentations
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation Muna Meky Impact Evaluation Cluster, AFTRL Slides by Paul J.
Advertisements

The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Choosing the level of randomization
Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
How to randomize? povertyactionlab.org. Quick Review: Why Randomize Choosing the Sample Frame Choosing the Unit of Randomization Options How to Choose.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Impact Evaluation Click to edit Master title style Click to edit Master subtitle style Impact Evaluation World Bank InstituteHuman Development Network.
BUSINESS AND FINANCIAL LITERACY FOR YOUNG ENTREPRENEURS: EVIDENCE FROM BOSNIA-HERZEGOVINA Miriam Bruhn and Bilal Zia (World Bank, DECFP)
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Making Impact Evaluations Happen World Bank Operational Experience 6 th European Conference on Evaluation of Cohesion Policy 30 November 2009 Warsaw Joost.
EXPERIMENTS AND OBSERVATIONAL STUDIES Chance Hofmann and Nick Quigley
Practical Sampling for Impact Evaluations Cyrus Samii, Columbia University.
Impact Evaluation Toolbox Gautam Rao University of California, Berkeley * ** Presentation credit: Temina Madon.
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
1 Randomization in Practice. Unit of randomization Randomizing at the individual level Randomizing at the group level –School –Community / village –Health.
Measuring Impact: Experiments
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Mattea Stein Quasi Experimental Methods.
AADAPT Workshop South Asia Goa, December 17-21, 2009 Nandini Krishnan 1.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Making the Most out of Discontinuities Florence.
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Private Sector Development Capacity-Building.
Why Use Randomized Evaluation? Presentation by Shawn Cole, Harvard Business School and J-PAL Presenter: Felipe Barrera-Osorio, World Bank 1 APEIE Workshop.
Beyond surveys: the research frontier moves to the use of administrative data to evaluate R&D grants Oliver Herrmann Ministry of Business, Innovation.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
Impact Evaluation Designs for Male Circumcision Sandi McCoy University of California, Berkeley Male Circumcision Evaluation Workshop and Operations Meeting.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Why Use Randomized Evaluation? Isabel Beltran, World Bank.
Applying impact evaluation tools A hypothetical fertilizer project.
Impact Evaluation for Evidence-Based Policy Making
What is randomization and how does it solve the causality problem? 2.3.
Measuring Impact 1 Non-experimental methods 2 Experiments
Development Impact Evaluation in Finance and Private Sector 1.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
TRANSLATING RESEARCH INTO ACTION What and How to randomize? July 9, 2011 Dhaka Raymond Guiteras, Assistant Professor University of Maryland povertyactionlab.org.
CHOOSING THE LEVEL OF RANDOMIZATION. Unit of Randomization: Individual?
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Laura Chioda.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 Steps in Implementing an Impact.
Country: Mauritius Manufacturing and Services Development and Competitiveness Project CROSS-COUNTRY WORKSHOP FOR IMPACT EVALUATIONS IN FINANCE AND PRIVATE.
Implementing an impact evaluation under constraints Emanuela Galasso (DECRG) Prem Learning Week May 2 nd, 2006.
Randomized Assignment Difference-in-Differences
Impact Evaluation Using Impact Evaluation for Results Based Policy Making Arianna Legovini Impact Evaluation Cluster, AFTRL Slides by Paul J. Gertler &
Overview of evaluation of SME policy – Why and How.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Social Experimentation & Randomized Evaluations Hélène Giacobino Director J-PAL Europe DG EMPLOI, Brussells,Nov 2011 World Bank Bratislawa December 2011.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Common Pitfalls in Randomized Evaluations Jenny C. Aker Tufts University.
Development Impact Evaluation Initiative Innovations in investment climate reforms Paris, Nov 13, 2012 In collaboration with the Investment Climate Global.
IMPACT EVALUATION PBAF 526 Class 5, October 31, 2011.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Why Use Randomized Evaluation?
Using Randomized Evaluations to Improve Policy
Measuring Results and Impact Evaluation: From Promises into Evidence
Using Randomized Evaluations to Improve Policy
Development Impact Evaluation in Finance and Private Sector
Impact Evaluation Toolbox
Implementation Challenges
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Impact Evaluation Designs for Male Circumcision
Steps in Implementing an Impact Evaluation
Presentation transcript:

Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Using Randomized Evaluations to Improve Policy Isabel Beltran

Objective  To Identify the True Effect of a Program separate the impact of the program from other factors >> What is the causal effect of a program?  Need to find out what would have happened without the program Cannot observe the same person with and without the program >> Rely on counterfactual analysis (control group) 2

Correlation is not causation Higher profits Credit Use OR ? 2) ? 1) Higher profits Business Skills Credit Question: Does providing credit increase firm profits? Suppose we observe that firms with more credit also earn higher profits.

4 (+6) increase in gross operating margin Illustration: Credit Program (1) A credit program was offered in Why did operating margin increase?

5 (+) Impact of the program (+) Impact of other (external) factors Illustration: Credit Program (2)

6 (+) BIASED Measure of the program impact Illustration: Credit Program (Before-After)

Motivation  Hard to distinguish causation from correlation from statistical analysis of existing data However complex the statistics can only see that X (credit program) goes with Y (operating margin) Hard to correct for unobserved characteristics, like motivation/ability Motivation/ability may be the most important things to correct for  Selection bias a major issue for impact evaluation Projects started at specific times and places for particular reasons Participants may be selected or self-select into programs (eligibility criteria)  People who have access to credit are likely to be very different from the average entrepreneur, looking at their profits will give you a misleading impression of the benefits of credit 7

Motivation  Retrospective impact evaluation: Collecting data after the event you don’t know how participants and nonparticipants compared before the program started Have to try and disentangle why the project was implemented where and when it was, after the event  Prospective evaluation allows you to design the evaluation to answer the question you need to answer  Allows you to collect the data you will need 8

Experimental Design  All those in the study have the same chance of being in the treatment or comparison group  By design treatment and comparison have the same characteristics (observed and unobserved), on average Only difference is treatment  With large sample, all characteristics average out  Unbiased impact estimates 9

Options for Randomization Lottery (only some get the program) Random phase-in  Everyone gets it eventually Variation in treatment  Full coverage but with different options Encouragement design (when partial take up)  Everyone can get it, some encouraged to get it 10

Example in Private Sector Development  Lottery Lottery to receive new loans  Random phase-in (everyone gets it eventually) Some groups or individuals get credit each year  Variation in treatment Some get matching grant, others get credit, others get business development services etc.  Encouragement design Some entrepreneurs get home visit to explain loan product, others do not 11

Lottery among the qualified Must get the program Not suitable for the program Randomize who gets the program

Opportunities for Randomization (1)  Budget constraints  prevent full coverage Random assignment (lottery) is fair and transparent  Limited implementation capacity Randomized phase-in gives all the same chance to go first  No evidence on which alternative is best Random assignment to alternatives with equal ex ante chance of success 13

Opportunities for Randomization (2)  Take up of existing program is not complete Provide information or incentive for some to sign up (Randomize encouragement)  Pilot a new program Good opportunity to test design before scaling up  Operational changes to ongoing programs Good opportunity to test changes before scaling them up 14

Different levels you can randomize at Individual (firm owner) Firm Business Association Village level Regulatory jurisdiction/ administrative district 15 Women’s association Youth groups School level

Group or individual randomization?  If a program impacts a whole group usually randomize whole group to treatment or comparison  Easier to get big enough sample if randomize individuals Individual randomizationGroup randomization

Unit of Randomization  Randomizing at higher level sometimes necessary: Political constraints on differential treatment within community Practical constraints—confusing for one person to implement different versions Spillover effects may require higher level randomization  Randomizing at group level requires many groups because of within community correlation 17

Elements of an experimental design 18 Random assignment Treatment GroupControl Group Participants  Drop-outs Evaluation sample Potential participants TailorsFurniture Manufacturers Target population SMEs

External and Internal Validity (1)  External validity The sample is representative of the total population The results in the sample represent the results in the population We can apply the lessons to the whole population  Internal validity The estimated effect of the intervention/program on the evaluated population reflects the real impact on that population i.e. the intervention and comparison groups are comparable 19

External and Internal Validity (2)  An evaluation can have internal validity without external validity Example: a randomized evaluation of encouraging informal firms to register in urban area may not tell you much about impact of a similar program in rural areas.  An evaluation without internal validity, can’t have external validity If you don’t know whether a program works in one place, then you have learnt nothing about whether it works elsewhere. 20

Internal AND external validity 21 Randomization National Population Samples National Population

Internal validity ONLY 22 Stratification (some sub- group of the population) Randomization (treatment and control) Population Population stratum Samples of Population Stratum

23 Representative but biased: useless National Population Biased assignment USELESS! Randomization

Example: Credit Program, internal validity 24 Random assignment Sample of female entrepreneurs

Example: Credit program Basic sequence of tasks for the evaluation  Listing eligible firms ▪ E.g. SMEs with turnover below a certain threshold  Baseline data of firms  Random assignment to different treatments or to treatment and control  Project implementation  Follow-up survey 25

Efficacy & Effectiveness  Efficacy Proof of concept Smaller scale Pilot in ideal conditions  Effectiveness At scale Prevailing implementation arrangements -- “real life”  Higher or lower impact?  Higher or lower costs? 26

Advantages of “experiments” Clear and precise causal impact Relative to other methods:  Much easier to analyze  Cheaper (smaller sample sizes)  Easier to explain  More convincing to policymakers  Methodologically uncontroversial 27

What if there are constraints on randomization?  Budget constraints: randomize among the most in need  Roll-out capacity constraints: randomize who receives first… (or next, if you have already started)  Randomly promote the program to some and not to others …(participants make their own choices about adoption) 28

 Those who get/receive promotion or marketing are more likely to enroll  But who got promotion or marketing was determined randomly, so not correlated with other observables or non-observables Compare average outcomes of two groups: promoted/not promoted Effect of offering the encouragement (Intent-To- Treat) Effect of the intervention on the complier population (Local Average Treatment Effect) ▪ LATE= effect of offering program (ITT)/proportion of those who took it up

Assigned to treatment Assigned to control DifferenceImpact: Average treatment effect on the treated Non-treated Treated Proportion treated 100%0%100% Impact of assignment 100% Mean outcome Intent-to-treat estimate 23/100%=23 Average treatment on the treated

EncouragedNot encouraged DifferenceImpact: Average treatment effect on compliers Non-treated (did not take up program) Treated (did take up program) Proportion treated 70%30%40% Impact of encouragement 100% Outcome Intent-to-treat estimate 8/40%=20 Average treatment on the compliers

Common pitfalls to avoid Calculating sample size incorrectly Randomizing one district to treatment and one district to control and calculating sample size on number of people you interview Collecting data in treatment and control differently Counting those assigned to treatment who do not take up program as control—don’t undo your randomization!! 32

When is it really not possible?  The treatment already assigned and announced and no possibility for expansion of treatment  The program is over (retrospective)  Universal take-up already  Program is national and non excludable Freedom of the press, exchange rate policy (sometimes some components can be randomized)  Sample size is too small to make it worth it 33

Thank you Financial support from: Bank Netherlands Partnership Program (BNPP), Bovespa, CVM, Gender Action Plan (GAP), Belgium & Luxemburg Poverty Reduction Partnerships (BPRP/LPRP), Knowledge for Change Program (KCP), Russia Financial Literacy and Education Trust Fund (RTF), and the Trust Fund for Environmentally & Socially Sustainable Development (TFESSD), is gratefully acknowledged.