The brass standard? The scope for RCTs in impact evaluation Jan Willem Gunning VU University Amsterdam University of Sussex, December 5, 2014.

Slides:



Advertisements
Similar presentations
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Advertisements

1 Research in Micro Finance: Big Questions and How to Answer Them.
Mywish K. Maredia Michigan State University
Sample size issues & Trial Quality David Torgerson.
Chapter 11 What Works and What Doesn’t. Are Hospitals Good for You? From Angrist and Pischke, Mostly Harmless Econometrics.
The counterfactual logic for public policy evaluation Alberto Martini hard at first, natural later 1.
Raymond Martin What is Research? “A STUDIOUS ENQUIRY or examination especially a critical and exhaustive investigation or experimentation.
BUSINESS AND FINANCIAL LITERACY FOR YOUNG ENTREPRENEURS: EVIDENCE FROM BOSNIA-HERZEGOVINA Miriam Bruhn and Bilal Zia (World Bank, DECFP)
VIII Evaluation Conference ‘Methodological Developments and Challenges in UK Policy Evaluation’ Daniel Fujiwara Senior Economist Cabinet Office & London.
Econometric Modeling. How do we determine if econometric analysis is credible? Leamer (1983) focused on robustness: What is robustness? The sensitivity.
Public Policy & Evidence: How to discriminate, interpret and communicate scientific research to better inform society. Rachel Glennerster Executive Director.
Who are the participants? Creating a Quality Sample 47:269: Research Methods I Dr. Leonard March 22, 2010.
Statistics Micro Mini Threats to Your Experiment!
The Role of Financial System in Economic Growth Presented By: Saumil Nihalani.
The Use of Decision Analysis in Program Evaluation Farrokh Alemi, Ph.D.
TOOLS OF POSITIVE ANALYSIS
Evaluation Eyal Ophir CS 376 4/28/09. Readings Methodology Matters (McGrath, 1994) Practical Guide to Controlled Experiments on the Web (Kohavi et al.,
DevelopmentEconomics. Development Economics Introductionto.
PAI786: Urban Policy Class 2: Evaluating Social Programs.
Experimental Design The Gold Standard?.
TRADUIRE LA RECHERCHE EN ACTION Employment RCTs in France Bruno Crépon.
Study Design for Quantitative Evaluation ©2015 The Water Institute.
WELCOME THE 12 TH ANNUAL MARCES/MSDE CONFERENCE: Value Added Modeling and Growth Modeling with Particular Application to Teacher and School Effectiveness.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
Randomized Control Trials for Agriculture Pace Phillips, Innovations for Poverty Action
ECON ECON Health Economic Policy Lab Kem P. Krueger, Pharm.D., Ph.D. Anne Alexander, M.S., Ph.D. University of Wyoming.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Randomized Controlled Trials in Rural Finance: An Example from India Michael Faye and Sendhil Mullainathan Harvard University March 2007
Economic evaluation of drugs for rare diseases CENTRE FOR HEALTH ECONOMICS K Claxton, C McCabe, A Tsuchiya Centre for Health Economics and Department of.
Assessing the Distributional Impact of Social Programs The World Bank Public Expenditure Analysis and Manage Core Course Presented by: Dominique van de.
RiPPLE Theme 3 WSS, Livelihoods & Growth London November, 2007.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
SOCIOLOGICAL INVESTIGATION
ECON 3039 Labor Economics By Elliott Fan Economics, NTU Elliott Fan: Labor 2015 Fall Lecture 21.
Ch. 2 Tools of Positive Economics. Theoretical Tools of Public Finance theoretical tools The set of tools designed to understand the mechanics behind.
Chapter 3.1.  Observational Study: involves passive data collection (observe, record or measure but don’t interfere)  Experiment: ~Involves active data.
The Choice Between Fixed and Random Effects Models: Some Considerations For Educational Research Clarke, Crawford, Steele and Vignoles and funding from.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Randomised Controlled Trials: What, why and how? Pam Hanley 22 March 2013.
Fixed Effects Models Evaluation Research (8521) Prof. Jesse Lecy 1.
Economics 172 Issues in African Economic Development Lecture 21 April 11, 2006.
Applying impact evaluation tools A hypothetical fertilizer project.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Steps in Implementing an Impact Evaluation Nandini Krishnan.
Research Design Week 6 Part February 2011 PPAL 6200.
Randomized controlled trials and the evaluation of development programs Chris Elbers VU University and AIID 11 November 2015.
Sifting through the evidence Sarah Fradsham. Types of Evidence Primary Literature Observational studies Case Report Case Series Case Control Study Cohort.
Firm Size, Finance and Growth Thorsten Beck Asli Demirguc-Kunt Luc Laeven Ross Levine.
S-005 Types of research in education. Types of research A wide variety of approaches: –Theoretical studies –Summaries of studies Reviews of the literature.
Introduction to the discussion of the presentation by Y. L’Horty ASIRPA International WS 13 June 2012 ASIRPA Socio-economic analysis of the diversity of.
Public Finance and Public Policy Jonathan Gruber Third Edition Copyright © 2010 Worth Publishers 1 of 24 Copyright © 2010 Worth Publishers.
Children’s Emotional and Behavioral Problems and Their Parents’ Labor Supply Patrick Richard, Ph.D., M.A. Nicholas C. Petris Center on Health Markets and.
Experimental Evaluations Methods of Economic Investigation Lecture 4.
Improving Lives through Impact Evaluation Dr. Jyotsna (Jo) Puri Head of Evaluation Deputy Executive Director, 3ie Presentation at the IEO Inauguration.
Critically Appraising a Medical Journal Article
Measuring Results and Impact Evaluation: From Promises into Evidence
Evidence-based Medicine
UKES Annual Conference
Take-home quiz due! Get out materials for notes!
Impact evaluation: The quantitative methods with applications
Impact Evaluation Methods
Empirical Tools of Public Finance
1 Causal Inference Counterfactuals False Counterfactuals
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Class 2: Evaluating Social Programs
Sampling for Impact Evaluation -theory and application-
Class 2: Evaluating Social Programs
Principles of Science and Systems
Sample Sizes for IE Power Calculations.
Development Economics.
Presentation transcript:

The brass standard? The scope for RCTs in impact evaluation Jan Willem Gunning VU University Amsterdam University of Sussex, December 5, 2014

Assessing policy impact in development economics growing concern about endogeneity (growth regressions) scepticism about IV-estimates the example of medical research (testing drugs): RCTs But drugs testing is fundamentally different! extraordinary claims: development economis is RCTs

The randomista´s gold standard: only RCTs can identify causal effects don’t waste time on questions not suitable for RCTs RCT results have (unlimited) external validity Abhijit Banerjee en Esther Duflo

RCTs the gold standard? “ experiments have no special ability to produce more credible knowledge than other methods” Angus Deaton, JEL, 2010

Ignore the big questions? “Instead of discussing how best to fight diarrhea or dengue, many of the most vocal experts tend to be fixated on the “big questions”: What is the ultimate cause of poverty? How much faith should we place in free markets? Is democracy good for the poor? Does foreign aid have a role to play? And so on.” Banerjee en Duflo, Poor Economics, 2011

RCTs under fire.. not suitable for important questions Rodrik (2008), Ravallion (2012) doubts about external validity Banerjee and He (2008) vs. Banerjee and Duflo (2011) RCT may affect the control group Deaton (2010) upscaling changes the intervention Pritchett and Sandefur (2013), Bold et al. (2013) RCT misses selection on the gain Elbers and Gunning (2013)

Who benefits? selection on the gain assignment and treatment effect correlated program officer bases assignment in part on private information on treatment effect examples: NGO drinking water and sanitation program: local staff select program villages credit program: loan officer selects clients

RCTs in trouble..

Can RCTs deal with this? no, the problem is fundamental RCT (randomisation over beneficiaries) does not mimic the actual assignment process hence the RCT gives a correct answer to an irrelevant question randomising instead over program officers: loss of statistical power internal validity undermined: characteristics of program officers correlated with controls rehabilitation for observational data? such data became suspect because of endogeneity concerns.. but in this case they can do the trick.. if data from a reprentative sample: external validity assured

Estimating the total program effect (TPE) regression with observational data

Does it matter? example: impact evaluation of a health insurance program in Vietnam accounting for treatment heterogeneity doubles the health effect simple test: joint significance of the additional regressors in the TPE regression

Conclusion in development programs assignment often neither universal nor random, but correlated with treatment effects IV-regressions and RCTs then both wrong the impact evaluation literature focuses on an irrelevant parameter: Eβ the relevant parameter EβP can be estimated using observational rather than experimental data

Complication: P and X correlated examples: educational policies induce changes in parents’ contribution (Dercon); local governments reduce activities in response to central government programs (Deaton) unidirectional P caused by X: no adjustment unidirectional X caused by P: estimate TPE without the relevant X terms – this gives direct and indirect effect of P bidirectional case: if effect of P on X can be identified then indirect effect can be estimated and added to the direct effect 13