AADAPT Workshop South Asia Goa, India, December 17-21, 2009 Arianna Legovini Manager, Development Impact Evaluation Initiative The World Bank.

Slides:



Advertisements
Similar presentations
Armenias Millennium Challenge Account: Assessing Impacts Ken Fortson, MPR Ester Hakobyan, MCA Anahit Petrosyan, MCA Anu Rangarajan, MPR Rebecca Tunstall,
Advertisements

PAYING FOR PERFORMANCE In PUBLIC HEALTH: Opportunities and Obstacles Glen P. Mays, Ph.D., M.P.H. Department of Health Policy and Administration UAMS College.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation Muna Meky Impact Evaluation Cluster, AFTRL Slides by Paul J.
Centre for Strategy & Evaluation Services Evaluation of the SME Funding Schemes - summary European Agency for Safety and Health at Work.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Donald T. Simeon Caribbean Health Research Council
Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
Global Poverty Action Fund Community Partnership Window Funding Seminar January 2014 Global Poverty Action Fund Community Partnership Window Funding Seminar.
The 8-7 National Poverty Reduction Program in China: the National Strategy and its Impact Wang Sangui, Li Zhou, Ren Yanshun.
Goal Paper  Improve our understanding on whether business training can improve business practices and firm outcomes (sales, profits, investment) of poor.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
1 Randomization in Practice. Unit of randomization Randomizing at the individual level Randomizing at the group level –School –Community / village –Health.
Randomised controlled trials Peter John. Causation in policy evaluation Outcome Intervention Other agency actions External environment.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Nandini Krishnan Africa Impact Evaluation Initiative World Bank April 14, 2009.
Measuring Impact: Experiments
AADAPT Workshop South Asia Goa, December 17-21, 2009 Nandini Krishnan 1.
SEDA IMPACT EVALUATION WESTERN CAPE (SOUTH AFRICA) Varsha Harinath (the dti) Francisco Campos (World Bank) Finance and Private Sector Development IE Workshop.
Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)
Use of impact evaluation for operational research PREM Week 2007 Arianna Legovini, Africa Impact Evaluation Initiative (AFTRL)
PREM Week April 2009 PREM Week April 2009 Arianna Legovini Head, Development Impact Evaluation Initiative (DIME) World Bank Impact Evaluation for Real.
AADAPT Workshop South Asia Goa, December 17-21, 2009 Maria Isabel Beltran 1.
Epilepsy and WHO | 17 Oct |1 | WHO's six-point agenda The overarching health needs 1.Promoting development 2.Fostering health security The strategic.
Impact Evaluation for Real Time Decision Making Arianna Legovini Head, Development Impact Evaluation Initiative (DIME) World Bank.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Impact Evaluation of CDBC under Vietnam Second Northern Mountains Poverty Reduction Project Ms. Nguyen Thi Minh Nghia Mr. Le Hong Phong Mr. Tran Duy Hung.
Evaluating Ongoing Programs: A Chronological Perspective to Include Performance Measurement Summarized from Berk & Rossi’s Thinking About Program Evaluation,
Why Use Randomized Evaluation? Isabel Beltran, World Bank.
Africa Program for Impact Evaluation on HIV/AIDS (AIM-AIDS) Cape Town, March 2009 Workshop of the Africa Program for Impact Evaluation of HIV/AIDS.
Applying impact evaluation tools A hypothetical fertilizer project.
Impact Evaluation for Evidence-Based Policy Making
S Ethiopia Sustainable Land Management Project Impact Evaluation CROSS-COUNTRY WORKSHOP FOR IMPACT EVALUATIONS IN AGRICULTURE AND COMMUNITY DRIVEN DEVELOPMENT.
An introduction to Impact Evaluation and application to the Ethiopia NFSP Workshop on Approaches to Evaluating The Impact of the National Food Security.
AADAPT Workshop for Impact Evaluation in Agriculture and Rural Development Goa, India 2009 With generous support from Gender Action Plan.
Development Impact Evaluation in Finance and Private Sector 1.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Outline of the Presentation I.Context II.Key Questions III. How do Countries Rate? IV. Conclusion.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 Steps in Implementing an Impact.
Randomized Assignment Difference-in-Differences
Impact Evaluation of Urban Upgrading Programs Judy Baker, FEU November 19, 2007.
Pilot and Feasibility Studies NIHR Research Design Service Sam Norton, Liz Steed, Lauren Bell.
Agricultural (rain and irrigation) water management across landscape for sustainable intensification and smallholders resilience building.
Objective 1: To increase resilience of smallholder production systems Output -Integrated crop-livestock systems developed to improve productivity, profitability.
1 Ongoing Agriculture Projects in Nigeria and Malawi Ahmed Mushfiq Mobarak Yale University.
Ghana: Impact of the Productive Safety Nets Project on Agricultural Productivity Angela Dannson, MoFA Benjamin Botchway, MoLG Osman Gyasi, World Bank Markus.
What is Impact Evaluation … and How Do We Use It? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Common Pitfalls in Randomized Evaluations Jenny C. Aker Tufts University.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
Operational Aspects of Impact Evaluation
Impact Evaluation for Real Time Decision Making
Measuring Results and Impact Evaluation: From Promises into Evidence
Right-sized Evaluation
Seminario Avaliacão de Impacto em Agricultura e Desinvolvimento Rural
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 2. Session 6. Developing indicators.
Impact Evaluation Terms Of Reference
Implementation Issues Program roll-out
LIVESTOCK PRODUCTION AND PRODUCTIVITY
Institutionalizing the Use of Impact Evaluation
Workshop of the Africa Program for Impact Evaluation of HIV/AIDS
Development Impact Evaluation in Finance and Private Sector
Implementation Challenges
III. Practical Considerations in preparing a CIE
Sampling for Impact Evaluation -theory and application-
Steps in Implementing an Impact Evaluation
Sample Sizes for IE Power Calculations.
Steps in Implementing an Impact Evaluation
Presentation transcript:

AADAPT Workshop South Asia Goa, India, December 17-21, 2009 Arianna Legovini Manager, Development Impact Evaluation Initiative The World Bank

 To secure operational relevance of IE and change culture  Learning agenda must be set up bottom up  This requires capacity development for IE in implementing agencies  Some formal training  Mainly application and learning by doing by being part of the evaluation team  Objective  use impact evaluation as an internal and routine management tool  secure policy feedback

 A learning agenda is a set of questions of high policy and operational importance, the answers for which will:  Secure policy continuity  Improve effectiveness of programs  Buy approval of policy establishment, electorate How?  Know your sector  Involve all relevant parties to the discussion  Include evaluators to help structure the discussion and resulting analytical design  Identify program(s) with highest potential, political sensitivity or with high fiscal expenditures  Develop forward looking learning agenda  Include policy (what) and operational (how to) questions  Disseminate learning agenda to relevant parties (minister, constituencies)

 Question design-choices of program  Institutional arrangements, Delivery mechanisms, Packages, Pricing/incentive schemes  Use random trials to test alternatives  Measure effects on short term outcomes (as leading indicators or other things to come)  take up rates, use, adoption  Follow up data collection and analysis  months after exposure  Measure impact of alternative treatments on short term outcomes and identify “best”  Change program to adopt best alternative  Start over

 How much does the program deliver?  Is it cost-effective?  Use most rigorous method of evaluation possible  Focus on higher level outcomes  Agricultural productivity, health status, income  Measure impact of operation on stated objectives and a metric of common outcomes  One, two, three year horizon  Compare with results from other programs  Inform budget process and allocations

 Take next year decisions  Justify changes to the program  Negotiate your budget  Justify expansion

 Agriculture Sector priority: Increasing commercialization of agricultural products  Intervention: Grants for added value projects  Priority for learning:  What training is needed to help farmer associations succeed?  What level of subsidy is most cost effective?  For which product lines are the grant most effective?  Governance Sector priority: Improving local accountability  Intervention: Budgetary support conditional on participatory decision- making  Priority for learning:  What are the rules of the game that are most conducive to driving decisions toward public and away from private goods?  How much effort should be exerted to ensure participation of women, poor, etc?

 Determine scale: Large scale or pilot?  Universal scale with imperfect take-up: encouragement design  Universal scale with perfect take-up: difficult  Large scale with representative sample: more costly, more informative  Large scale with purposeful sample: less costly, good for first instance, may require more evaluation later  Small pilot (e.g., in two districts): easier to implement, not as informative, may need to use all beneficiaries  Some programs are too small to evaluate

 Identify eligibility criteria and targeting  Rolled out in communities/population/regions satisfying a certain criteria? Possibility to use regression discontinuities  Rolled out to a targeted high-potential or high-poverty population/areas?  Understand roll out (timing and geography)  Piloted in a sample of households, communities, or regions? Possibility for random assignment  Rolled out nationwide? In how many phases? How many villages per phase? Possibility for random phase in  Universal program with imperfect take up? Possibility for random encouragement  Investigate budget constraints Possibility for random assignment Each roll-out strategy yields distinct opportunities for impact evaluation

 Each question requires a separate design or “identification strategy”  For each question there will be a method that is best in the sense that it provides the most precise estimate and is operationally feasible  When unsure plan to use more than one method  Keep ethical consideration in mind:  No to deny benefits to something that we know works  Test interventions before scale up if we don’t know  Discuss design with authorizing environment

 Random encouragement: use random assignment of promotional activities to evaluate  Promotion activities on take-up  Grants and incentives on productivity (use random encouragement as an instrument to create exogenous variation in take up)  Randomize in the call for proposal pipeline:  Call for “expressions of interest”  Select twice+ as many “expressions of interest” than those you can fund  Randomly select half of them and ask them to submit full proposals (treatment)  The other half serve as control (either forever or until next call)

 Power calculations needed to determine size of sample to be used in the evaluation  What is the unit of intervention?  Sample increases the smaller the expected effects  If clustered (village, community) ▪ Power decreases with high (intracluster) correlation of outcomes—need more clusters ▪ Power mostly determined by the number of clusters not the number of observations within each cluster

 Cost items:  Capacity development  Project team time to manage IE  Analytical services for design and analysis  Field coordination  Data collection  Discussions and dissemination

 IE Team should include  Government (program manager, economist/statistician)  WB Project team (Task manager or substitute)  Research team (Lead researcher, coresearchers, field coordinator)  Data collection agency

 The smallest unit of assignment is the unit of intervention  Pension: individual  Health insurance: individual or household  Female president of local council: village  Assignment can be at higher aggregation level-but costly  Treatment units are assigned intervention and control units are not—create listing and records  Assignment must be discussed and explained to all parties who make decisions on implementation to avoid contamination later on

 Only collect baseline data AFTER impact evaluation design is ready  Impact evaluation design determines sample (use power calculations)  Random assignment: control and treatment are identical at baseline  Baseline strictly not needed for analysis, used to check for balance, reassign if needed; use if randomization is contaminated  Quasi-experimental methods:  Baseline essential  Baseline Analysis:  Informs project design and implementation; improve targeting

 IE team (not data collection agency) to  Design questionnaire and sample  Define terms of reference for data collection agency  Train enumerators  Conduct pilot  Supervise data collection

 Design questionnaire(s) and sample that meet monitoring and impact evaluation data needs  Final outcomes : yield, consumption, incomes  Intermediate outcomes we expect to change first: input use, technology adoption  Other outcomes that the intervention may affect: schooling, labor  Characteristics that might affect outcomes: farm size, household size, education  In short, outcomes of interest AND variables that help understand how the intervention affect different populations

 Who?  Bureau of Statistics: Integrate with existing data  Ministry concerned: Ministry of Agriculture/Water Resources/Rural Development  Private agency: Sometimes higher quality, more dependable

 Do treatment and control groups look similar at baseline?  If not, all is not lost!  Even in absence of perfect balance, can use baseline data to adjust analysis or re-assign PovertyFemale- headed households Number of children in household Formal sector job Treatment70%64%3.120% Control68%66%2.918% Significance-*--

 Intensive monitoring of roll-out to ensure evaluation is not compromised  What if treatment and control receive the benefits? ▪ Input vouchers randomly assigned to households, but rolled out to the entire community, or treatment households sell their vouchers to control households ▪ Is the evaluation is compromised? Needed to monitor!  What if all the control group receive some other benefit? ▪ NGO targets control communities to receive vouchers ▪ Changes evaluation: comparison between your program and the NGO program.

 Collect follow-up data with the same sample and questionnaire as baseline data  Appropriate intervals  Consider how long it should take for outcomes to change  One year or at next harvest ▪ Provide initial outcomes ▪ Adjust program if needed  Two years: Changes in longer term outcomes?  After end of program: Do effects endure? ▪ What happens once the input voucher program has phased out?

 Randomization: Simply compare average outcomes for treatment and comparison  Other methods: Do econometric analysis required taking in consideration assumptions to estimate impact of program  Combination of methods:  Random Encouragement and IV  Matching with difference-in-difference

 Are the effects statistically significant?  Basic statistical test tells whether differences are due to the program or to noisy data  Are they significant in real terms?  If the input voucher scheme costs a million dollars and has positive effect but it’s tiny, may not be worthwhile  Are they sustainable?  If input use falls to pre-program levels when the intervention ends, the program is not financially sustainable in its current form

 Are you thinking about this only now???  Discuss what are the policy implications of the results  What actions should be taken  How to present them to authorizing environment to justify changes/budget/scale up?  Talk to policy-maker and disseminate to wider audience  If no one knows about it, it won’t make a difference  Make sure the information gets into the right policy discussions  Real time discussions  Workshops  Reports  Policy briefs

 Identify next learning opportunity  Test variations  Alternate subsidy amounts  Alternate packages of inputs  Alternate implementation and targeting mechanisms: Government extension workers or input dealers? Beneficiary selection?  Test other interventions to affect same outcomes  Matching grants for technology adoption  Training in use of improved technologies  Improving access to markets and providing complementary infrastructure to increase the share of marketed output