Randomized Assignment Difference-in-Differences

Slides:



Advertisements
Similar presentations
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation Muna Meky Impact Evaluation Cluster, AFTRL Slides by Paul J.
Advertisements

The World Bank Human Development Network Spanish Impact Evaluation Fund.
Designing an impact evaluation: Randomization, statistical power, and some more fun…
The World Bank Human Development Network Spanish Impact Evaluation Fund.
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Impact Evaluation Click to edit Master title style Click to edit Master subtitle style Impact Evaluation World Bank InstituteHuman Development Network.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Propensity Score Matching Lava Timsina Kristina Rabarison CPH Doctoral Seminar Fall 2012.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
© Institute for Fiscal Studies The role of evaluation in social research: current perspectives and new developments Lorraine Dearden, Institute of Education.
Impact Evaluation Session VII Sampling and Power Jishnu Das November 2006.
Development Impact Evaluation Initiative innovations & solutions in infrastructure, agriculture & environment naivasha, april 23-27, 2011 in collaboration.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Welfare Reform and Lone Parents Employment in the UK Paul Gregg and Susan Harkness.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Matching Estimators Methods of Economic Investigation Lecture 11.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Session III Regression discontinuity (RD) Christel Vermeersch LCSHD November 2006.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Impact Evaluation Designs for Male Circumcision Sandi McCoy University of California, Berkeley Male Circumcision Evaluation Workshop and Operations Meeting.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Public Policy Analysis ECON 3386 Anant Nyshadham.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
1 2 3 Economist Sociologist Project Officer. Evaluating Anti-Poverty Programs: Concepts and Methods Martin Ravallion Development Research Group, World.
What is randomization and how does it solve the causality problem? 2.3.
Measuring Impact 1 Non-experimental methods 2 Experiments
Christel M. J. Vermeersch November 2006 Session V Instrumental Variables.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Using Propensity Score Matching in Observational Services Research Neal Wallace, Ph.D. Portland State University February
Implementing an impact evaluation under constraints Emanuela Galasso (DECRG) Prem Learning Week May 2 nd, 2006.
WBI WORKSHOP Randomization and Impact evaluation.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Effects of migration and remittances on poverty and inequality A comparison between Burkina Faso, Kenya, Nigeria, Senegal, South Africa, and Uganda Y.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
The Evaluation Problem Alexander Spermann, University of Freiburg 1 The Fundamental Evaluation Problem and its Solution SS 2009.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
Alexander Spermann University of Freiburg, SS 2008 Matching and DiD 1 Overview of non- experimental approaches: Matching and Difference in Difference Estimators.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
The Evaluation Problem Alexander Spermann, University of Freiburg, 2007/ The Fundamental Evaluation Problem and its Solution.
General belief that roads are good for development & living standards
Sec 9C – Logistic Regression and Propensity scores
An introduction to Impact Evaluation
Quasi-Experimental Methods
Impact Evaluation Methods
Explanation of slide: Logos, to show while the audience arrive.
Instrumental Variables and Two Stage Least Squares
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Instrumental Variables and Two Stage Least Squares
Impact Evaluation Methods
1 Causal Inference Counterfactuals False Counterfactuals
Matching Methods & Propensity Scores
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Impact Evaluation Designs for Male Circumcision
Explanation of slide: Logos, to show while the audience arrive.
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Module 3: Impact Evaluation for TTLs
Presentation transcript:

Randomized Assignment Difference-in-Differences 2 Randomized Assignment Randomized Promotion Discontinuity Design Difference-in-Differences Diff-in-Diff IE Methods Toolbox Matching P-Score matching

Choosing your IE method(s) Key information you will need for identifying the right method for your program: Prospective/Retrospective Evaluation? Eligibility rules and criteria? Poverty targeting? Geographic targeting? Roll-out plan (pipeline)? Is the number of eligible units larger than available resources at a given point in time? Budget and capacity constraints? Excess demand for program? Etc.

Choosing your IE method(s) Choose the best possible design given the operational context: Best Design Best comparison group you can find + least operational risk Have we controlled for everything? Internal validity Good comparison group External validity Local versus global treatment effect Evaluation results apply to population we’re interested in Is the result valid for everyone?

Randomized Assignment Difference-in-Differences 2 Randomized Assignment Randomized Promotion Discontinuity Design Difference-in-Differences Diff-in-Diff IE Methods Toolbox Matching P-Score matching

Matching Idea How? Issue? For each treated unit pick up the best comparison unit (match) from another data source. How? Matches are selected on the basis of similarities in observed characteristics. Issue? If there are unobservable characteristics and those unobservables influence participation: Selection bias!

Propensity-Score Matching (PSM) Comparison Group: non-participants with same observable characteristics as participants. In practice, it is very hard. There may be many important characteristics! Match on the basis of the “propensity score”, Solution proposed by Rosenbaum and Rubin: Compute everyone’s probability of participating, based on their observable characteristics. Choose matches that have the same probability of participation as the treatments. See appendix 2.

Density of propensity scores Non-Participants Participants Common Support 1 Propensity Score

Case 7: Progresa Matching (P-Score) Baseline Characteristics Estimated Coefficient Probit Regression, Prob Enrolled=1 Head’s age (years) -0.022** Spouse’s age (years) -0.017** Head’s education (years) -0.059** Spouse’s education (years) -0.03** Head is female=1 -0.067 Indigenous=1 0.345** Number of household members 0.216** Dirt floor=1 0.676** Bathroom=1 -0.197** Hectares of Land -0.042** Distance to Hospital (km) 0.001* Constant 0.664** Note: If the effect is statistically significant at the 1% significance level, we label the estimated impact with 2 stars (**).

Case 7: Progresa Common Support Density: Pr (Enrolled) Density: Pr (Enrolled) Density: Pr (Enrolled) Pr (Enrolled)

Case 7: Progresa Matching (P-Score) Estimated Impact on Consumption (Y) Multivariate Linear Regression 7.06+ Note: If the effect is statistically significant at the 1% significance level, we label the estimated impact with 2 stars (**). If significant at 10% level, we label impact with +

! Keep in Mind Matching Matching requires large samples and good quality data. Ex-post matching is risky: If there is no baseline, be careful! matching on endogenous ex-post variables gives bad results. Matching at baseline can be very useful: Know the assignment rule and match based on it combine with other techniques (i.e. diff-in-diff)

Progresa Policy Recommendation? Impact of Progresa on Consumption (Y) Case 1: Before & After 34.28** Case 2: Enrolled & Not Enrolled -4.15 Case 3: Randomized Assignment 29.75** Case 4: Randomized Promotion 30.4** Case 5: Discontinuity Design 30.58** Case 6: Differences in Differences 25.53** Case 7: Matching 7.06+ Note: If the effect is statistically significant at the 1% significance level, we label the estimated impact with 2 stars (**). If significant at 10% level, we label impact with +

Progresa Policy Recommendation? Impact of Progresa on Consumption (Y) Case 1: Before & After 34.28** Case 2: Enrolled & Not Enrolled -4.15 Case 3: Randomized Assignment 29.75** Case 4: Randomized Promotion 30.4** Case 5: Discontinuity Design 30.58** Case 6: Differences in Differences 25.53** Case 7: Matching 7.06+ Note: If the effect is statistically significant at the 1% significance level, we label the estimated impact with 2 stars (**). If significant at 10% level, we label impact with +

Randomized Assignment Difference-in-Differences 2 Randomized Assignment Randomized Promotion Combinations of methods Discontinuity Design Difference-in-Differences Diff-in-Diff IE Methods Toolbox Matching P-Score matching

(Eligibility Cut-off) (No Eligibility Cut-off) Choosing your method Targeted (Eligibility Cut-off) Universal (No Eligibility Cut-off) Limited Resources (Never Able to Achieve Scale) Fully Resourced (Able to Achieve Scale) Phased Implementation Over Time Randomized Assignment RDD Randomized Assignment (roll-out) Matching with DiD Immediate Implementation Random Assignment Random Promotion

Remember “ The objective of impact evaluation is to estimate the causal effect or impact of a program on outcomes of interest.

“ Remember To estimate impact, we need to estimate the counterfactual. what would have happened in the absence of the program and use comparison or control groups.

Remember “ We have a toolbox with 5 methods to identify good comparison groups.

Remember “ Choose the best evaluation method that is feasible in the program’s operational context.

Appendix 1 Two Stage Least Squares (2SLS) 2nd and 3rd day IE methods for Policy Makers Appendix 1 Two Stage Least Squares (2SLS) Model with endogenous Treatment (T): Stage 1: Regress endogenous variable on the IV (Z) and other exogenous regressors: Calculate predicted value for each observation: T hat

Appendix 1 Two Stage Least Squares (2SLS) 2nd and 3rd day IE methods for Policy Makers Appendix 1 Two Stage Least Squares (2SLS) Stage 2: Regress outcome y on predicted variable (and other exogenous variables): Need to correct Standard Errors (they are based on T hat rather than T) In practice just use STATA – ivreg. Intuition: T has been “cleaned” of its correlation with ε.

Appendix 2 Steps in Propensity Score Matching Representative & highly comparables survey of non-participants and participants. Pool the two samples and estimated a logit (or probit) model of program participation. Restrict samples to assure common support (important source of bias in observational studies) For each participant find a sample of non-participants that have similar propensity scores Compare the outcome indicators. The difference is the estimate of the gain due to the program for that observation. Calculate the mean of these individual gains to obtain the average overall gain.