Impact Evaluation for Evidence-Based Policy Making

Slides:



Advertisements
Similar presentations
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation Muna Meky Impact Evaluation Cluster, AFTRL Slides by Paul J.
Advertisements

The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Explanation of slide: Logos, to show while the audience arrive.
Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
Mywish K. Maredia Michigan State University
Impact Evaluation Click to edit Master title style Click to edit Master subtitle style Impact Evaluation World Bank InstituteHuman Development Network.
What could go wrong? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop Africa Program for Education.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
Pricing the right to education The cost of reaching new targets by 2030 Aaron Benavot Director, EFA Global Monitoring Report Launch Event, Results for.
Global Poverty Action Fund Community Partnership Window Funding Seminar January 2014 Global Poverty Action Fund Community Partnership Window Funding Seminar.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Making Impact Evaluations Happen World Bank Operational Experience 6 th European Conference on Evaluation of Cohesion Policy 30 November 2009 Warsaw Joost.
Types of Evaluation.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
Measuring Impact: Experiments
AADAPT Workshop South Asia Goa, December 17-21, 2009 Nandini Krishnan 1.
1 What are Monitoring and Evaluation? How do we think about M&E in the context of the LAM Project?
The National Development Plan, Iraq 6 July 2010 “Developing Objectives & Indicators for Strategic Planning” Khaled Ehsan and Helen Olafsdottir UNDP Iraq.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
MAINSTREAMING MONITORING AND EVALUATION IN EDUCATION Can education be effectively managed without an M & E system in place?
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Private involvement in education: Measuring Impacts Felipe Barrera-Osorio HDN Education Public-Private Partnerships in Education, Washington, DC, March.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
Impact Evaluation Designs for Male Circumcision Sandi McCoy University of California, Berkeley Male Circumcision Evaluation Workshop and Operations Meeting.
Impact Evaluation for Real Time Decision Making Arianna Legovini Head, Development Impact Evaluation Initiative (DIME) World Bank.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Why Use Randomized Evaluation? Isabel Beltran, World Bank.
Africa Program for Impact Evaluation on HIV/AIDS (AIM-AIDS) Cape Town, March 2009 Workshop of the Africa Program for Impact Evaluation of HIV/AIDS.
Applying impact evaluation tools A hypothetical fertilizer project.
An introduction to Impact Evaluation and application to the Ethiopia NFSP Workshop on Approaches to Evaluating The Impact of the National Food Security.
AADAPT Workshop for Impact Evaluation in Agriculture and Rural Development Goa, India 2009 With generous support from Gender Action Plan.
There and Back Again: An Impact Evaluator’s Tale Paul Gertler University of California, Berkeley July, 2015.
Measuring Impact 1 Non-experimental methods 2 Experiments
Managing Public Budget to Facilitate Economic Growth and Reduce Poverty Public Expenditure Analysis & Management Staff Training Course May , 2001.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
From Monitoring Through Evaluation To Impact Assessment The case of NHDRs and vulnerable groups Andrey Ivanov Human Development Adviser, Bratislava RSC.
Randomized Assignment Difference-in-Differences
Impact Evaluation Using Impact Evaluation for Results Based Policy Making Arianna Legovini Impact Evaluation Cluster, AFTRL Slides by Paul J. Gertler &
Social Experimentation & Randomized Evaluations Hélène Giacobino Director J-PAL Europe DG EMPLOI, Brussells,Nov 2011 World Bank Bratislawa December 2011.
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Using Randomized Evaluations to Improve.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Monitoring and evaluation Objectives of the Session  To Define Monitoring, impact assessment and Evaluation. (commonly know as M&E)  To know why Monitoring.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Monitoring and evaluation 16 July 2009 Michael Samson UNICEF/ IDS Course on Social Protection.
1 An introduction to Impact Evaluation (IE) for HIV/AIDS Programs March 12, 2009 Cape Town Léandre Bassolé ACTafrica, The World Bank.
Impact Evaluation Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead, Africa Impact Evaluation Initiative AFTRL.
Introduction to Impact Evaluation The Motivation Emmanuel Skoufias The World Bank PRMPR PREM Learning Week: April 21-22, 2008.
Operational Aspects of Impact Evaluation
Measuring Results and Impact Evaluation: From Promises into Evidence
Right-sized Evaluation
Explanation of slide: Logos, to show while the audience arrive.
Institutionalizing the Use of Impact Evaluation
Development Impact Evaluation in Finance and Private Sector
1 Causal Inference Counterfactuals False Counterfactuals
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Evaluating Impacts: An Overview of Quantitative Methods
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Impact Evaluation Designs for Male Circumcision
Operational Aspects of Impact Evaluation
Steps in Implementing an Impact Evaluation
Presentation transcript:

Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead, Africa Impact Evaluation Initiative

Answer Three Questions Why is evaluation valuable? What makes a good impact evaluation? How to implement evaluation?

IE Answers: How do we turn this teacher…

…into this teacher?

Why Evaluate? Need evidence on what works Allocate limited budget Fiscal accountability Improve program/policy overtime Operational research Managing by results Information key to sustainability Negotiating budgets Informing constituents and managing press Informing donors

Traditional M&E and Impact Evaluation monitoring to track implementation efficiency (input-output) impact evaluation to measure effectiveness (output-outcome) BEHAVIOR MONITOR EFFICIENCY INPUTS OUTPUTS OUTCOMES EVALUATE EFFECTIVENESS $$$

Question types and methods Process Evaluation / Monitoring: Is program being implemented efficiently? Is program targeting the right population? Are outcomes moving in the right direction? Descriptive analysis Impact Evaluation: What was the effect of the program on outcomes? How would outcomes change under alternative program designs? Does the program impact people differently (e.g. females, poor, minorities)? Is the program cost-effective? Causal analysis

Which can be answered by traditional M&E and which by IE? Are books being delivered as planned? Does de-worming increase school attendance? What is the correlation between enrollment and school quality? Does the decentralized school management lead to an increase in learning achievement?

Types of Impact Evaluation Efficacy: Proof of Concept Pilot under ideal conditions Effectiveness: At scale Normal circumstances & capabilities Lower or higher impact? Higher or lower costs?

So, use impact evaluation to…. Test innovations Scale up what works (e.g. de-worming) Cut/change what does not (e.g. HIV counseling) Measure effectiveness of programs (e.g. JTPA ) Find best tactics to e.g. change people’s behavior (e.g. come to the clinic) Manage expectations e.g. PROGRESA/OPORTUNIDADES (Mexico) Transition across presidential terms Expansion to 5 million households Change in benefits Battle with the press

Next question please What makes a good impact evaluation? Why is evaluation valuable? What makes a good impact evaluation? How to implement evaluation?

Assessing impact examples How much do girl scholarships increase school enrollment? What is the level of beneficiary’s learning achievement with program compared to without program? Compare same individual with & without programs at the same point in time Never observe same individual with and without program at same point in time

Solving the evaluation problem Counterfactual: what would have happened without the program Need to estimate counterfactual i.e. find a control or comparison group Counterfactual Criteria Treated & counterfactual groups have identical initial characteristics on average, Only reason for the difference in outcomes is due to the intervention

2 “Counterfeit” Counterfactuals Before and after: Same individual before the treatment Non-Participants: Those who choose not to enroll in program Those who were not offered the program

Before and After Example Food Aid Compare mortality before and after Find increase in mortality Did the program fail? “Before” normal year, but “after” famine year Cannot separate (identify) effect of food aid from effect of drought

Before and After Compare Y before and after intervention B Before-after counterfactual A-B Estimated impact Control for time varying factors C True A-C True impact A-B is under-estimated Y Before After C A B B t-1 t Time Treatment

Non-Participants…. Compare non-participants to participants Counterfactual: non-participant outcomes Problem: why did they not participate?

Exercise: Why participants and non-participants differ? Access to school Poorer Unmet demand More organized community Achievement Poverty Gender Children who come to school and children who do not? Communities that applied for funds for a new classroom and communities that did not? Children who received scholarships and children who did not?

Literacy program example Treatment offered Who signs up? Those who are illiterate Have lower education than those who do not sign up Educated people are a poor estimate of counterfactual

What's wrong? Selection bias: People choose to participate for specific reasons Many times reasons are directly related to the outcome of interest Cannot separately identify impact of the program from these other factors/reasons

Program placement example Government offers school inputs program to schools with low infrastructure Compare achievement in schools offered program to achievement in schools not offered Program targeted based on lack of inputs, so Treatments have low achievement Counterfactuals have high achievement Cannot separately identify program impact from school targeting criteria

Need to know… Why some get program and others do not How some get into treatment and other in control group If reasons correlated with outcome cannot identify/separate program impact from other explanations of differences in outcomes The process by which data is generated

Possible Solutions… Guarantee comparability of treatment and control groups ONLY remaining difference is intervention In this workshop we will consider Experimental design/randomization Quasi-experiments Regression Discontinuity Double differences Instrumental Variables

These solutions all involve… Randomization Give all equal chance of being in control or treatment groups Guarantees that all factors/characteristics will be on average equal between groups Only difference is the intervention If not, need transparent & observable criteria for who is offered program

How to implement evaluation? The Last Question Why is evaluation valuable? What makes a good impact evaluation? How to implement evaluation?

Implementation Issues Political economy Policy context Finding a good control Retrospective versus prospective designs Making the design compatible with operations Ethical Issues Relationship to “results” monitoring

Political Economy What is the policy purpose? In USA test innovations to national policy, defend budget In RSA answer electorate In Mexico allocate budget to poverty programs In IDA country pressure to demonstrate aid effectiveness and scale up In poor country hard constraints and ambitious targets: how to reach those targets?

Evidence culture and incentives for change Cultural shift From retrospective evaluation Look back and judge To prospective evaluation Decide what need to learn Experiment with alternatives Measure and inform Adopt better alternatives overtime Change in incentives Rewards for changing programs that do not work Rewards for generating knowledge Separating job performance from knowledge generation

The Policy Context Address policy-relevant questions: What policy questions need answers? What outcomes answer those questions? What indicators measures outcomes? How much of a change in the outcomes would determine success? Example: teacher performance-based pay Scale up pilot? Criteria: Need at least a 10% increase in test scores with no change in unit costs

Opportunities for good designs Use opportunities to generate good control groups Most programs cannot deliver benefits to all those eligible Budgetary limitations: Eligible who get it are potential treatments Eligible who do not are potential controls Logistical limitations: Those who go first are potential treatments Those who go later are potential controls

Who gets the program? Who goes first? Eligibility criteria Roll out Are benefits targeted? How are they targeted? Can we rank eligible's priority? Are measures good enough for fine rankings? Who goes first? Roll out Equal chance to go first, second, third?

Ethical Considerations Do not delay benefits: Rollout based on budget/administrative constraints Equity: equally deserving beneficiaries deserve an equal chance of going first Transparent & accountable method Give everyone eligible an equal chance If rank based on some criteria, then criteria should be quantitative and public

Retrospective Designs Hard to find good control groups Must live with arbitrary or unobservable allocation rules Administrative data good enough to reflect program was implemented as described Need pre-intervention baseline survey On both controls and treatments With covariates to control for initial differences Without baseline difficult to use quasi-experimental methods

Manage for results Retrospective evaluation cannot be used to manage for results Use resources wisely: do prospective evaluation design Better methods More tailored policy questions Precise estimates Timely feedback and program changes Improve results on the ground

Monitoring Systems Projects/programs regularly collect data for management purposes Typical content Lists of beneficiaries Distribution of benefits Expenditures Outputs Ongoing process evaluation Information is needed for impact evaluation

Evaluation uses administrative information to: Verify who is beneficiary When started What benefits were actually delivered Necessary condition for program to have an impact: benefits need to get to targeted beneficiaries

Improve use of administrative data for IE Program monitoring data usually only collected in areas where active Collect baseline for control areas as well Very cost-effective as little need for additional special surveys Add a couple of outcome indicators Most IE’s use only monitoring data

Overall Messages Impact evaluation useful for Validating program design Adjusting program structure Communicating to finance ministry & civil society A good evaluation design requires estimating the counterfactual What would have happened to beneficiaries if had not received the program Need to know all reasons why beneficiaries got program & others did not

Design Messages Address policy questions Stakeholder buy-in Interesting is what government needs and will use Stakeholder buy-in Easiest to use prospective designs Good monitoring systems & administrative data can improve IE and lower costs