Different Methods of Impact Evaluation

Slides:



Advertisements
Similar presentations
5.1 Rules for Exponents Review of Bases and Exponents Zero Exponents
Advertisements

There is a pattern for factoring trinomials of this form, when c
1 Research Designs For Evaluating Disease Management Program Effectiveness Ariel Linden, Dr.P.H., M.S. President, Linden Consulting Group Disease Management.
Multiplication X 1 1 x 1 = 1 2 x 1 = 2 3 x 1 = 3 4 x 1 = 4 5 x 1 = 5 6 x 1 = 6 7 x 1 = 7 8 x 1 = 8 9 x 1 = 9 10 x 1 = x 1 = x 1 = 12 X 2 1.
Math Expressions How does it all work together…?.
Non-randomized Medical Device Clinical Studies: A Regulatory Perspective Sep. 16, 2005 Lilly Yue, Ph.D.* CDRH, FDA, Rockville MD * No official support.
The Application of Propensity Score Analysis to Non-randomized Medical Device Clinical Studies: A Regulatory Perspective Lilly Yue, Ph.D.* CDRH, FDA,
Combining Like Terms. Only combine terms that are exactly the same!! Whats the same mean? –If numbers have a variable, then you can combine only ones.
Impact Evaluation Methods: Causal Inference
/4/2010 Box and Whisker Plots Objective: Learn how to read and draw box and whisker plots Starter: Order these numbers.
1 Eloise E. Kaizar The Ohio State University Combining Information From Randomized and Observational Data: A Simulation Study June 5, 2008 Joel Greenhouse.
Impact analysis and counterfactuals in practise: the case of Structural Funds support for enterprise Gerhard Untiedt GEFRA-Münster,Germany Conference:
1 First steps in practice Daniel Mouqué Evaluation Unit DG REGIO.
The logic of C ounterfactual I mpact E valuation 1.
Entry task 1) What can be concluded from the following pattern? = 15 = 3 × = 20 = 4 × = 25 =
1 1  1 =.
1  1 =.
Slides prepared by JOHN LOUCKS St. Edward’s University.
The basics for simulations
13.1 Theoretical Probability
Chapter 4: Basic Estimation Techniques
(This presentation may be used for instructional purposes)
PP Test Review Sections 6-1 to 6-6
Regression with Panel Data
Control in Experimental Designs. Control u Key element of experimental and quasi- experimental designs –Subjects (one group or several) on which no variable.
Data Analysis 53 Given the two histograms below, which of the following statements are true?
N ON -E XPERIMENTAL M ETHODS Shwetlena Sabarwal (thanks to Markus Goldstein for the slides)
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Adding Up In Chunks.
Designing Experiments
Mediation: Multiple Variables David A. Kenny. 2 Mediation Webinars Four Steps Indirect Effect Causal Assumptions.
Let’s take a 15 minute break Please be back on time.
Advantages and limitations of non- and quasi-experimental methods Module 2.2.
12 System of Linear Equations Case Study
& dding ubtracting ractions.
Class 18 – Thursday, Nov. 11 Omitted Variables Bias
Experimental Design and Analysis of Variance
Where do data come from and Why we don’t (always) trust statisticians.
Review of Identifying Causal Effects Methods of Economic Investigation Lecture 13.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Chapter 13 Notes Observational Studies and Experimental Design
TRANSLATING RESEARCH INTO ACTION What is Randomized Evaluation? Why Randomize? J-PAL South Asia, April 29, 2011.
Matching Estimators Methods of Economic Investigation Lecture 11.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Chapter 3.1.  Observational Study: involves passive data collection (observe, record or measure but don’t interfere)  Experiment: ~Involves active data.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Producing Data: Experiments BPS - 5th Ed. Chapter 9 1.
The Evaluation Problem Alexander Spermann, University of Freiburg 1 The Fundamental Evaluation Problem and its Solution SS 2009.
Alexander Spermann University of Freiburg, SS 2008 Matching and DiD 1 Overview of non- experimental approaches: Matching and Difference in Difference Estimators.
Differences-in-Differences
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Methods of Economic Investigation Lecture 12
Impact Evaluation Methods
Matching Methods & Propensity Scores
Ch. 13. Pooled Cross Sections Across Time: Simple Panel Data.
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Cases. Simple Regression Linear Multiple Regression.
Ch. 13. Pooled Cross Sections Across Time: Simple Panel Data.
Presentation transcript:

Different Methods of Impact Evaluation

How to measure impact? Assessing causality Event 2 (Effects) Event 1 Impact of program = outcome 1 – outcome 2 In practice, we compare two groups, one of which benefited from the program, the other one did not Event 2 (Effects) Outcome 1 Event 1 e.g. Education program/ treatment Causes Event 2 occurs if and only if Event 1 occurred before Event 1 No Education Program Event 2 Outcome 2 Causes

Constructing the counterfactual what would have happened in the absence of the program? … … for the people who benefitted from the program: we don’t observe it thus, impact evaluation will have to mimic it Counterfactual is constructed by selecting a group not affected by the program – this is the main challenge of impact evaluation Methods differ by the way the counterfactual is constructed

Non-Experimental Methods Differences between outcomes for treated and non-treated (in our ex: reached / not) are the sum of Inherent differences: “SELECTION” BIAS TREATMENT EFFECT: that we want to isolate from 1 The essence of non-experimental methods is to find a way to redress the selection bias ex-post. The critical objective of impact evaluation is to establish a credible comparison group – a group of individuals who in the absence of the program would have had outcomes similar to those who were exposed to the program. However, in reality it is generally the case that individuals who participate in a program and those who were not are different: programs are placed in specific areas (for example, poorer or richer areas) individuals are screened for participation in the program (for example, on the basis of poverty or on the basis of their motivation) and, in addition, the decision to participate is often voluntary. For all of these reasons, those who were not exposed to a program are often not a good comparison group for those who were, and any differences between the groups can be attributed to two factors: pre-existing differences (selection bias) and the impact of the program. Since we have no reliable way to estimate the size of the selection bias, we typically cannot decompose the overall difference into a treatment effect and a bias term. Selection bias disappears in the case of randomization

Non-Experimental Methods Simple Difference Multivariate (Multiple) Regression “Difference in difference” (Multiple Regression with Panel Data) Matching Randomization/RCT’s (advanced topic covered separately) The critical objective of impact evaluation is to establish a credible comparison group – a group of individuals who in the absence of the program would have had outcomes similar to those who were exposed to the program. However, in reality it is generally the case that individuals who participate in a program and those who were not are different: programs are placed in specific areas (for example, poorer or richer areas) individuals are screened for participation in the program (for example, on the basis of poverty or on the basis of their motivation) and, in addition, the decision to participate is often voluntary. For all of these reasons, those who were not exposed to a program are often not a good comparison group for those who were, and any differences between the groups can be attributed to two factors: pre-existing differences (selection bias) and the impact of the program. Since we have no reliable way to estimate the size of the selection bias, we typically cannot decompose the overall difference into a treatment effect and a bias term. Selection bias disappears in the case of randomization

Simple difference Simple difference is a first measure – why is it not sufficient? There may be differences between the two groups (age, location, gender, initial endowment, bargaining power) So we may want to control for these differences

Multi-variate regression Suppose we can observe these differences. Age composition of the group, initial infrastructure in the school, level of education, … We can include all these variables in a regression: Y = a.T + b.Age + c.Infr + d.Edu +… A regression provides the linear combination of observable variables that best “mimics” the outcome Each coefficient represents the effect of each variable Will give us the effect of the treatment everything else being equal, or more exactly every other observable characteristics being equal

Multi-variate regression Problems of the regression You may want to include many many variables, to control for as many characteristics as possible Problem of sample size (degrees of freedom) More important: do you have measures of everything? Bargaining power, Pro-activeness, Intrinsic motivation, hopelessness There are unobservables

Panel data Simple difference: before/after Counterfactual = same group before the program Can we trust this? Assumption = would have remained the same Ex: Police project Double difference Control for the situation before the program Ex: Group 1 = Treatment Group; Group 2 = Control Group 2006: Group 1 : 30 Group 2 : 60 2009: Group 1 : 50 (+66%) Group 2 : 90 (+50%) Effect = +16% Assumption: they would have grown at the same pace Not sure…

Matching We compare pairs of 2 individuals for which the values taken by ALL variables are the same

Matching Variation: Propensity Score Matching: all the variables do not need to be exactly the same, but you look for individuals which have the same “profile” Problems This matching method requires a big dataset: find pairs on a sufficient number of variables What about unobservables?

An example Case study: US Congress elections, 2002 60 000 phone calls to potential voters to encourage them to vote; 25 000 reached Outcome: did they actually go vote? 1st method: compare the 25 000 (reached) vs. the 35 000 (not reached) 2nd method: introduce co-variates in a regression 3rd method: introduce baseline data (vote in 1998) 4th method: do a matching

1st comparison: we suspect a selection bias Reached Not Reached Difference Female 56.2% 53.8% 2.4 pp* Newly Regist. 7.3% 9.6% -2.3 pp* From Iowa 54.7% 46.7% 8.0 pp* Voted in 2000 71.7% 63.3% 8.3 pp* Voted in 1998 46.6% 37.6% 9.0 pp* Selection Bias: differences in observable / unobservable characteristics → differences in outcome not due to the treatment 13 13

Non-Experimental Methods Estimated Impact 1 – Simple Difference 10.8 pp * 2 – Multiple regression 6.1 pp * 3 – Multiple regression with panel data (diff-in-diff) 4.5 pp * 4 – Matching 2.8 pp * 5 – Randomized Experiment 0.4 pp

For more details, please read Case Study: “Get out the vote For more details, please read Case Study: “Get out the vote? Do phone calls encourage voting” under References