Quasi Experimental Methods I

Slides:



Advertisements
Similar presentations
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Advertisements

N ON -E XPERIMENTAL M ETHODS Shwetlena Sabarwal (thanks to Markus Goldstein for the slides)
Advantages and limitations of non- and quasi-experimental methods Module 2.2.
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
1 COMM 301: Empirical Research in Communication Lecture 10 Kwan M Lee.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Experimental Design making causal inferences. Causal and Effect The IV precedes the DV in time The IV precedes the DV in time The IV and DV are correlated.
Correlation AND EXPERIMENTAL DESIGN
Jeff Beard Lisa Helma David Parrish Start Presentation.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
Development Impact Evaluation Initiative innovations & solutions in infrastructure, agriculture & environment naivasha, april 23-27, 2011 in collaboration.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Mattea Stein Quasi Experimental Methods.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Experimental Design making causal inferences Richard Lambert, Ph.D.
Matching Estimators Methods of Economic Investigation Lecture 11.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
Study Session Experimental Design. 1. Which of the following is true regarding the difference between an observational study and and an experiment? a)
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
What is randomization and how does it solve the causality problem? 2.3.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Chapter Eight: Quantitative Methods
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Social Experimentation & Randomized Evaluations Hélène Giacobino Director J-PAL Europe DG EMPLOI, Brussells,Nov 2011 World Bank Bratislawa December 2011.
How Psychologists Do Research Chapter 2. How Psychologists Do Research What makes psychological research scientific? Research Methods Descriptive studies.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Randomization.
Measuring causal impact 2.1. What is impact? The impact of a program is the difference in outcomes caused by the program It is the difference between.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
1 An introduction to Impact Evaluation (IE) for HIV/AIDS Programs March 12, 2009 Cape Town Léandre Bassolé ACTafrica, The World Bank.
Differences-in-Differences
Lurking inferential monsters
The Science of Social Psychology
Measuring Results and Impact Evaluation: From Promises into Evidence
CLINICAL PROTOCOL DEVELOPMENT
Quasi Experimental Methods I
An introduction to Impact Evaluation
Quasi-Experimental Methods
Quasi-Experimental Methods
Chapter Eight: Quantitative Methods
2 independent Groups Graziano & Raulin (1997).
An introduction to Impact Evaluation
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
ESF EVALUATION PARTNERSHIP MEETING Bernhard Boockmann / Helmut Apel
Experimental Design.
Development Impact Evaluation in Finance and Private Sector
Establishing the Direction of the Relationship
Impact Evaluation Methods
Experimental Design.
Scientific Method Steps
Matching Methods & Propensity Scores
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Scientific Method Steps
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Randomization This presentation draws on previous presentations by Muna Meky, Arianna Legovini, Jed Friedman, David Evans and Sebastian Martinez.
Statistical Data Analysis
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Psych 231: Research Methods in Psychology
CHAPTER 10 Comparing Two Populations or Groups
Psych 231: Research Methods in Psychology
Psych 231: Research Methods in Psychology
Presentation transcript:

Quasi Experimental Methods I Florence Kondylis Non-Experimental Methods Quasi Experimental Methods I This presentation draws on previous presentations by Markus Goldstein, Leandre Bassole, Florence Kondylis and Alberto Martini

What we know so far Aim: We want to isolate the causal effect of our interventions on our outcomes of interest Use rigorous evaluation methods to answer our operational questions Randomizing the assignment to treatment is the “gold standard” methodology (simple, precise, cheap) What if we really, really (really??) cannot use it?! >> Where it makes sense, resort to non-experimental methods

When does it make sense? Can we find a plausible counterfactual? Natural experiment? Every non-experimental method is associated with a set of assumptions The stronger the assumptions, the more doubtful our measure of the causal effect Question our assumptions Reality check, resort to common sense!

Example: Fertilizer Voucher Program Principal Objective Increase maize production Intervention Fertilizer vouchers distribution Non-random assignment Target group Maize producers, land over 1 Ha & under 5 Ha Main result indicator Maize yield

Illustration: Fertilizer Voucher Program (1) (+) Impact of the program (+) Impact of external factors 6

Illustration: Fertilizer Voucher Program (2) (+) BIASED Measure of the program impact If no baseline data and no experimental design, can abandon the idea of measuring true impact (unless we have a natural experiment) “Before-After” doesn’t deliver results we can believe in! 7

Illustration: Fertilizer Voucher Program (3) « Before» difference btwn participants and nonparticipants « After » difference btwn participants and non-participants >> What’s the impact of our intervention? 8

Difference-in-Differences Identification Strategy (1) Counterfactual: 2 Formulations that say the same thing Non-participants’ maize yield after the intervention, accounting for the “before” difference between participants/nonparticipants (the initial gap between groups) Participants’ maize yield before the intervention, accounting for the “before/after” difference for nonparticipants (the influence of external factors) 1 and 2 are equivalent

Difference-in-Differences Identification Strategy (2) Underlying assumption: Without the intervention, maize yield for participants and non participants’ would have followed the same trend >> Graphic intuition coming…

Data -- Example 1 1.3 1.9 0.6 1.4 0.8 0.7 0.5 -0.2 Average maize yield (T / Ha) 2007 2008 Difference (2007-2008) Participants (P) 1.3 1.9 0.6 Non-participants (NP) 1.4 0.8 Difference (P-NP) 0.7 0.5 -0.2

Data -- Example 1 1.3 1.9 0.6 1.4 0.8 0.7 0.5 -0.2 Average maize yield (T / Ha) 2007 2008 Difference (2007-2008) Participants (P) 1.3 1.9 0.6 Non-participants (NP) 1.4 0.8 Difference (P-NP) 0.7 0.5 -0.2

Impact = (P2008-P2007) -(NP2008-NP2007) = 0.6 – 0.8 = -0.2 P2008-P2007=0.6 NP2008-NP2007=0.8

Impact = (P-NP)2008-(P-NP)2007 = 0.5 - 0.7 = -0.2 P-NP2008=0.5 P-NP2007=0.7

Assumption of same trend: Graphic Implication Impact=-0.2

Summary Negative Impact: Assumption of same trend very strong Very counter-intuitive: Increased input use should not decrease yield once external factors are accounted for! Assumption of same trend very strong 2 groups were, in 2007, producing at very different levels Question the underlying assumption of same trend! When possible, test assumption of same trend with data from previous years

Questioning the Assumption of same trend: Use pre-pr0gram data >> Reject counterfactual assumption of same trends !

Data – Example 2 Average maize yield (T / Ha) 2007 2008 Difference (2007-2008) Participants (P) 1.5 2.1 0.6 Non-participants (NP) 0.5 0.7 0.2 Difference (P-NP) 1.0 1.4 0.4

Impact = (P2008-P2007) -(NP2008-NP2007) = 0.6 – 0.2 = + 0.4 NP08-NP07=0.2

Assumption of same trend: Graphic Implication Impact = +0.4

Conclusion Positive Impact: More intuitive Is the assumption of same trend reasonable? Still need to question the counterfactual assumption of same trends ! Use data from previous years

Questioning the Assumption of same trend: Use pre-pr0gram data >>Seems reasonable to accept counterfactual assumption of same trend ?!

Caveats (1) Assuming same trend is often problematic No data to test the assumption Even if trends are similar the previous year… Where they always similar (or are we lucky)? More importantly, will they always be similar? Example: Other project intervenes in our nonparticipant villages…

Caveats (2) What to do? >> Be descriptive! Check similarity in observable characteristics If not similar along observables, chances are trends will differ in unpredictable ways >> Still, we cannot check what we cannot see… And unobservable characteristics might matter more than observable (ability, motivation, patience, etc)

Matching Method + Difference-in-Differences (1) Match participants with non-participants on the basis of observable characteristics Counterfactual: Matched comparison group Each program participant is paired with one or more similar non-participant(s) based on observable characteristics >> On average, participants and nonparticipants share the same observable characteristics (by construction) Estimate the effect of our intervention by using difference-in-differences

Matching Method (2) Underlying counterfactual assumptions After matching, there are no differences between participants and nonparticipants in terms of unobservable characteristics AND/OR Unobservable characteristics do not affect the assignment to the treatment, nor the outcomes of interest

How do we do it? Design a control group by establishing close matches in terms of observable characteristics Carefully select variables along which to match participants to their control group So that we only retain Treatment Group: Participants that could find a match Comparison Group: Non-participants similar enough to the participants >> We trim out a portion of our treatment group!

Implications In most cases, we cannot match everyone Example Need to understand who is left out Example Matched Individuals Portion of treatment group trimmed out Nonparticipants Participants Score Wealth

Conclusion (1) Advantage of the matching method Does not require randomization

Conclusion (2) Disadvantages: Underlying counterfactual assumption is not plausible in all contexts, hard to test Use common sense, be descriptive Requires very high quality data: Need to control for all factors that influence program placement/outcome of choice Requires significantly large sample size to generate comparison group Cannot always match everyone…

Summary Randomized-Controlled-Trials require minimal assumptions and procure intuitive estimates (sample means!) Non-experimental methods require assumptions that must be carefully tested More data-intensive Not always testable Get creative: Mix-and-match types of methods! Adress relevant questions with relevant techniques

Thank you Financial support from Is gratefully acknowledged