Impact evaluation: The quantitative methods with applications

Slides:



Advertisements
Similar presentations
Evaluating Anti-Poverty Programs Part 1: Concepts and Methods Martin Ravallion Development Research Group, World Bank.
Advertisements

Impact analysis and counterfactuals in practise: the case of Structural Funds support for enterprise Gerhard Untiedt GEFRA-Münster,Germany Conference:
Different Methods of Impact Evaluation
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Evaluation of the impact of the Natural Forest Protection Programme on rural household incomes Katrina Mullan Department of Land Economy University of.
Propensity Score Matching Lava Timsina Kristina Rabarison CPH Doctoral Seminar Fall 2012.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Non Experimental Design in Education Ummul Ruthbah.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Welfare Reform and Lone Parents Employment in the UK Paul Gregg and Susan Harkness.
Evaluating Job Training Programs: What have we learned? Haeil Jung and Maureen Pirog School of Public and Environmental Affairs Indiana University Bloomington.
Evaluating the Efficacy of the Research Initiative for Scientific Enhancement (RISE) by using Propensity Scores to Identify a Matched Comparison Group.
Credit Scoring of Bank-affiliated Captive Finance Companies Gabriela Pásztorová CERGE-EI Bratislava Economic Meeting 8 June 2012.
Propensity Score Matching for Causal Inference: Possibilities, Limitations, and an Example sean f. reardon MAPSS colloquium March 6, 2007.
Public Policy Analysis ECON 3386 Anant Nyshadham.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
1 2 3 Economist Sociologist Project Officer. Evaluating Anti-Poverty Programs: Concepts and Methods Martin Ravallion Development Research Group, World.
Causal Structure, Endogeneity, and the Missing Data Problem in Modeling the Impact of Information and Communication Technology Use on Society Tuesday,
Lorraine Dearden Director of ADMIN Node Institute of Education
Using Propensity Score Matching in Observational Services Research Neal Wallace, Ph.D. Portland State University February
WBI WORKSHOP Randomization and Impact evaluation.
Randomized Assignment Difference-in-Differences
REBECCA M. RYAN, PH.D. GEORGETOWN UNIVERSITY ANNA D. JOHNSON, M.P.A. TEACHERS COLLEGE, COLUMBIA UNIVERSITY ANNUAL MEETING OF THE CHILD CARE POLICY RESEARCH.
MATCHING Eva Hromádková, Applied Econometrics JEM007, IES Lecture 4.
The Evaluation Problem Alexander Spermann, University of Freiburg 1 The Fundamental Evaluation Problem and its Solution SS 2009.
Alexander Spermann University of Freiburg, SS 2008 Matching and DiD 1 Overview of non- experimental approaches: Matching and Difference in Difference Estimators.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
Impact Evaluation Methods Regression Discontinuity Design and Difference in Differences Slides by Paul J. Gertler & Sebastian Martinez.
The Evaluation Problem Alexander Spermann, University of Freiburg, 2007/ The Fundamental Evaluation Problem and its Solution.
Looking for statistical twins
Kotchikpa Gabriel Lawin Lota Dabio Tamini
ECON 4009 Labor Economics 2017 Fall By Elliott Fan Economics, NTU
Module 4 Panel Data Jonathan Haughton June 2017 Measuring poverty
Module 5 Inference with Panel Data
Measuring Results and Impact Evaluation: From Promises into Evidence
Constructing Propensity score weighted and matched Samples Stacey L
General belief that roads are good for development & living standards
Pooling Cross Sections across Time: Simple Panel Data Methods
Propensity Score Matching
An introduction to Impact Evaluation
Impact Evaluation Methods
An introduction to Impact Evaluation
Explanation of slide: Logos, to show while the audience arrive.
Quasi-Experimental Methods
An introduction to Impact Evaluation
Pooling Cross Sections across Time: Simple Panel Data Methods
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Methods of Economic Investigation Lecture 12
Explanation of slide: Logos, to show while the audience arrive.
Impact Evaluation Methods
Impact Evaluation Methods
Matching Methods & Propensity Scores
Ch. 13. Pooled Cross Sections Across Time: Simple Panel Data.
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
The European Statistical Training Programme (ESTP)
Chapter: 9: Propensity scores
Analysing RWE for HTA: Challenges, methods and critique
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Positive analysis in public finance
Module 3: Impact Evaluation for TTLs
Ch. 13. Pooled Cross Sections Across Time: Simple Panel Data.
Alternative Scenarios and Related Techniques
Presentation transcript:

Impact evaluation: The quantitative methods with applications Evaluating the Impact of Projects and Programs Beijing, China, April 10-14, 2006 Shahid Khandker, WBI

5 Non-experimental Methods to Construct Data on Counterfactual Matching Propensity-score matching Difference-in-difference Matched double difference Instrumental variables

1. Matching Matched comparators identify counterfactual. Match participants to non-participants from a larger survey; The matches are chosen on the basis of similarities in observed characteristics; This assumes no selection bias based on unobservable heterogeneity; Validity of matching methods depends heavily on data quality.

2. Propensity-score Matching (PSM) Propensity-score matching match on the basis of the probability of participation. Ideally we would match on the entire vector X of observed characteristics. However, this is practically impossible. X could be huge. Rosenbaum and Rubin: match on the basis of the propensity score = This assumes that participation is independent of outcomes given X. If no bias given X then no bias given P(X).

Steps in Score Matching Representative, highly comparable, surveys of the non-participants and participants; (ii) Pool the 2 samples and estimate a logit/ probit model of program participation: Predicted values are “propensity scores.” (iii) Restrict samples to assure common support; (iv) Failure of common support is an important source of bias in observational studies (Heckman et al.).

Density of scores for participants

Density of scores for non-participants

Density of scores for non-participants

(v) For each participant find a sample of non-participants that have similar propensity scores; (vi) Compare the outcome indicators. The difference is the estimate of the gain due to the program for that observation; (vii) Calculate the mean of these individual gains to obtain the average overall gain. -Various weighting schemes.

The Mean Impact Estimator

How does PSM compare to an experiment? PSM is the observational analogue of an experiment in which placement is independent of outcomes; The difference is that a pure experiment does not require the untestable assumption of independence conditional on observables; Thus PSM requires good data.

How does PSM perform relative to other methods? In comparison with results of a randomized experiment on a US training program, PSM gave a good approximation (Heckman et al.; Dehejia and Wahba); Better than the non-experimental regression-based methods studied by Lalonde for the same program.

3. Difference-in-difference Observed changes over time for non-participants provide the counterfactual for participants. Collect baseline data on non-participants and (probable) participants before the program; Compare with data after the program; Subtract the two differences, or use a regression with a dummy variable for participant; This allows for selection bias but it must be time-invariant and additive.

Selection Bias Selection bias

DD requires that the bias is additive and time-invariant

Method fails if the comparison group is on a different trajectory

Outcome Indicator = impact (“gain”); Where: = counterfactual; = comparison group

Difference-In-Difference (i) if change over time for comparison group reveals counterfactual, and (ii) if baseline is uncontaminated by the program,

4. Matched double difference Matching helps control for bias in diff-in-diff. Score match participants and non-participants based on observed characteristics in baseline; Then do a double difference; This deals with observable heterogeneity in initial conditions that can influence subsequent changes over time.

5. Instrumental Variables Identifying exogenous variation using a 3rd variable Outcome regression: D = 0,1 is our program – not random “Instrument” (Z) influences participation, but does not affect outcomes given participation (the “exclusion restriction”). This identifies the exogenous variation in outcomes due to the program. Treatment regression: 5. Instrumental Variables

Reduced-form outcome regression: where and Instrumental variables (two-stage least squares) estimator of impact:

However, IVE is only a ‘local’ effect IVE identifies the effect for those induced to switch by the instrument “local average effect.” Suppose Z takes 2 values. Then the effect of the program is: Care in extrapolating to the whole population; Valid instruments can be difficult to find; Exclusion restrictions are often questionable.

Selected References Ravallion, Martin. “The Mystery of Vanishing Benefits” The World Bank Economic Review. Volume 15. No. 1. pp. 115-140. Washington D.C.: The World Bank, 2001. Khandker, Shahidur R. “Micro-finance and Poverty: Evidence Using Panel Data from Bangladesh” The World Bank Economic Review, October, 2005, PP: 263-286. Khandker, Shahidur R. “Fighting Poverty with Micro-credit: Experience in Bangladesh” New York, NY: Oxford University Press, 1998. Pitt, Mark M. and Shahidur R. Khandker. “The Impact of Group-Based Credit Programs on Poor Households in Bangladesh: Does the Gender of Participants Matter?” Journal of Political Economy 106 (October): 958-96, 1998.