Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.

Slides:



Advertisements
Similar presentations
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation Muna Meky Impact Evaluation Cluster, AFTRL Slides by Paul J.
Advertisements

The World Bank Human Development Network Spanish Impact Evaluation Fund.
An Overview Lori Beaman, PhD RWJF Scholar in Health Policy UC Berkeley
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Advantages and limitations of non- and quasi-experimental methods Module 2.2.
Mywish K. Maredia Michigan State University
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
Random Assignment Experiments
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Experimental Design making causal inferences. Causal and Effect The IV precedes the DV in time The IV precedes the DV in time The IV and DV are correlated.
Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
PAI786: Urban Policy Class 2: Evaluating Social Programs.
Impact Evaluation in the Real World One non-experimental design for evaluating behavioral HIV prevention campaigns.
Health Programme Evaluation by Propensity Score Matching: Accounting for Treatment Intensity and Health Externalities with an Application to Brazil (HEDG.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Welfare Reform and Lone Parents Employment in the UK Paul Gregg and Susan Harkness.
Correlational Research Chapter Fifteen Bring Schraw et al.
Matching Estimators Methods of Economic Investigation Lecture 11.
Beyond surveys: the research frontier moves to the use of administrative data to evaluate R&D grants Oliver Herrmann Ministry of Business, Innovation.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
Public Policy Analysis ECON 3386 Anant Nyshadham.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Measuring Impact 1 Non-experimental methods 2 Experiments
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Using Propensity Score Matching in Observational Services Research Neal Wallace, Ph.D. Portland State University February
Implementing an impact evaluation under constraints Emanuela Galasso (DECRG) Prem Learning Week May 2 nd, 2006.
WBI WORKSHOP Randomization and Impact evaluation.
Randomized Assignment Difference-in-Differences
1 Validating Ex Ante Impact Evaluation Models: An Example from Mexico Francisco H.G. Ferreira Phillippe G. Leite Emmanuel Skoufias The World Bank PREM.
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Public Finance and Public Policy Jonathan Gruber Third Edition Copyright © 2010 Worth Publishers 1 of 24 Copyright © 2010 Worth Publishers.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
1 The Training Benefits Program – A Methodological Exposition To: The Research Coordination Committee By: Jonathan Adam Lind Date: 04/01/16.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
Alexander Spermann University of Freiburg, SS 2008 Matching and DiD 1 Overview of non- experimental approaches: Matching and Difference in Difference Estimators.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
Patricia Gonzalez, OSEP June 14, The purpose of annual performance reporting is to demonstrate that IDEA funds are being used to improve or benefit.
Looking for statistical twins
Quasi Experimental Methods I
Constructing Propensity score weighted and matched Samples Stacey L
General belief that roads are good for development & living standards
Quasi Experimental Methods I
Propensity Score Matching
An introduction to Impact Evaluation
Quasi-Experimental Methods
Impact Evaluation Methods
Quasi-Experimental Methods
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Methods of Economic Investigation Lecture 12
Impact Evaluation Methods
Impact Evaluation Methods
Matching Methods & Propensity Scores
Implementation Challenges
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Module 3: Impact Evaluation for TTLs
Sample Sizes for IE Power Calculations.
Presentation transcript:

Matching Methods

Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline survey or time invariant characteristics  The matches are selected on the basis of similarities in observed characteristics  This assumes no selection bias based on unobserved characteristics Take the ITN Example from Yesterday: Households who were more concerned about malaria also took other preventative actions All such differences must be in the data in order for the match to produce a valid estimate of project impacts

Propensity-Score Matching (PSM) Propensity score matching: match treated and untreated observations on the estimated probability of being treated (propensity score). Most commonly used.  Match on the basis of the propensity score P(X) = Pr (d=1|X) D indicates participation in project Instead of attempting to create a match for each participant with exactly the same value of X, we can instead match on the probability of participation.

PSM: Key Assumptions  Key assumption: participation is independent of outcomes conditional on X i This is false if there are unobserved outcomes affecting participation  Enables matching not just at the mean but balances the distribution of observed characteristics across treatment and control

Density 0 1 Propensity score Region of common support Density of scores for participants High probability of participating given X Density of scores for non- participants

Steps in Score Matching 1. Need representative and comparable data for both treatment and comparison groups 2. Use a logit (or other discrete choice model) to estimate program participations as a function of observable characteristics 3. Use predicted values from logit to generate propensity score p(x i ) for all treatment and comparison group members

Calculating Impact using PSM 4. Match Pairs:  Restrict sample to common support (as in Figure)  Need to determine a tolerance limit: how different can control individuals or villages be and still be a match?  Nearest neighbors, nonlinear matching, multiple matches 5. Once matches are made, we can calculate impact by comparing the means of outcomes across participants and their matched pairs

PSM vs Randomization  Randomization does not require the untestable assumption of independence conditional on observables  PSM requires large samples and good data: 1. Ideally, the same data source is used for participants and non-participants 2. Participants and non-participants have access to similar institutions and markets, and 3. The data include X variables capable of identifying program participation and outcomes.

Lessons on Matching Methods  Typically used when neither randomization, RD or other quasi experimental options are not possible Case 1: no baseline. Can do ex-post matching Dangers of ex-post matching:  Matching on variables that change due to participation (i.e., endogenous)  What are some variables that won’t change?  Matching helps control only for OBSERVABLE differences, not unobservable differences

More Lessons on Matching Methods  Matching becomes much better in combination with other techniques, such as: Exploiting baseline data for matching and using difference-in-difference strategy If an assignment rule exists for project, can match on this rule  Need good quality data Common support can be a problem if two groups are very different

Case Study: Piped Water in India  Jalan and Ravaillion (2003): Impact of piped water for children’s health in rural India  Research questions of interest include: 1. Is a child less vulnerable to diarrhoeal disease if he/she lives in a HH with access to piped water? 2. Do children in poor, or poorly educated, HH have smaller health gains from piped water? 3. Does income matter independently of parental education?

Piped Water: the IE Design  Classic problem for infrastructure programs: randomization is generally not an option (although randomization in timing may be possible in other contexts)  The challenge: observable and unobservable differences across households with piped water and those without What are differences for such households in Nigeria?  Jalan and Ravallion use cross-sectional data nationally representative survey on 33,000 rural HH from 1765 villages

PSM in Practice  To estimate the propensity score, authors used:  Village level characteristics Including: Village size, amount of irrigated land, schools, infrastructure (bus stop, railway station)  Household variables Including: Ethnicity / caste / religion, asset ownership (bicycle, radio, thresher), educational background of HH members  Are there variables which can not be included? Only using cross-section, so no variables influenced by project

Piped Water: Behavioral Considerations  IE is designed to estimate not only impact of piped water but to look at how benefits vary across group  There is therefore a behavioral component: poor households may be less able to benefit from piped water b/c they do not properly store water  With this in mind, Are there any key variables missing?

Potential Unobserved Factors  The behavioral factors – importance put on sanitation and behavioral inputs – are also likely correlated with whether a HH has piped water  However, there are no behavioral variables in data: water storage, soap usage, latrines These are unobserved factors NOT included in propensity score

Piped Water: Impacts  Disease prevalence among those with piped water would be 21% higher without it  Gains from piped water exploited more by wealthier households and households with more educated mothers  Even find counterintuitive result for low income, illiterate HH: piped water is associated with higher diarrhea prevalence

DesignWhen to useAdvantagesDisadvantages Randomization  Whenever feasible  When there is variation at the individual or community level  Gold standard  Most powerful  Not always feasible  Not always ethical Randomized Encouragement Design  When an intervention is universally implemented  Provides exogenous variation for a subset of beneficiaries  Only looks at sub- group of sample  Power of encouragement design only known ex post Regression Discontinuity  If an intervention has a clear, sharp assignment rule  Project beneficiaries often must qualify through established criteria  Only look at sub- group of sample  Assignment rule in practice often not implemented strictly Difference-in- Differences  If two groups are growing at similar rates  Baseline and follow- up data are available  Eliminates fixed differences not related to treatment  Can be biased if trends change  Ideally have 2 pre- intervention periods of data Matching  When other methods are not possible  Overcomes observed differences between treatment and comparison  Assumes no unobserved differences (often implausible)