Public Policy & Evidence: How to discriminate, interpret and communicate scientific research to better inform society. Rachel Glennerster Executive Director.

Slides:



Advertisements
Similar presentations
The Role of Pilots Alex Bryson Policy Studies Institute ESRC Methods Festival, St Catherines College, Oxford University, 1st July 2004.
Advertisements

The World Bank Human Development Network Spanish Impact Evaluation Fund.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Advantages and limitations of non- and quasi-experimental methods Module 2.2.
1 Examples of Fixed-Effect Models. 2 Almond et al. Babies born w/ low birth weight(< 2500 grams) are more prone to –Die early in life –Have health problems.
1 Almond et al. Babies born w/ low birth weight(< 2500 grams) are more prone to – Die early in life – Have health problems later in life – Educational.
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
Drawing policy lessons from RCTs: When to generalize and when not to Rachel Glennerster J-PAL and IGC London School of Economics, June 3, 2014.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Valuing and Evaluating the Liberal Arts in the Research University The Liberal Arts & Sciences in the Research University Today: Histories, Challenges,
Chapter 11 What Works and What Doesn’t. Are Hospitals Good for You? From Angrist and Pischke, Mostly Harmless Econometrics.
Benefits and limits of randomization 2.4. Tailoring the evaluation to the question Advantage: answer the specific question well – We design our evaluation.
The counterfactual logic for public policy evaluation Alberto Martini hard at first, natural later 1.
Copyright EM LYON Par accord du CFC Cession et reproduction interdites Research in Entrepreneurship- The problem of unobserved heterogeneity Frédéric Delmar.
Who is the subject in a CRT? Rachel Glennerster Executive Director, J-PAL Department of Economics, MIT PRIM&R November 8, 2013.
Pre-analysis plans Module 8.3. Recap on statistics If we find a result is significant at the 5% level, what does this mean? – there is a 5% or less probability.
Differences-in-Differences
The World Bank Human Development Network Spanish Impact Evaluation Fund.
PHSSR IG CyberSeminar Introductory Remarks Bryan Dowd Division of Health Policy and Management School of Public Health University of Minnesota.
Correlation AND EXPERIMENTAL DESIGN
Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.
Non-Experimental designs: Developmental designs & Small-N designs
Non-Experimental designs: Developmental designs & Small-N designs
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Non-Experimental designs
Copyright © 2013 Wolters Kluwer Health | Lippincott Williams & Wilkins Statistical Methods for Health Care Research Chapter 1 Using Research and Statistics.
The Impacts of Microfinance: A Randomized Evaluation of Spandana’s program in Hyderabad June 8, 2009 Abhijit Banerjee Esther Duflo Rachel Glennerster Cynthia.
Postgraduate Course 6. Evidence based management: What is the best available evidence?
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, 2009 AIM-CDD Using Randomized Evaluations.
What do we know about improving health in Punjab? And some lessons from India Dr. Jeffrey Hammer Princeton University IGC – CDPR Seminar, Islamabad February.
Chapter 1 - Introduction & Research Methods What is development?
Research Methods in Human Sexuality
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
Global Workshop on Development Impact Evaluation in Finance and Private Sector Rio de Janeiro, June 6-10, 2011 Mattea Stein Quasi Experimental Methods.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
S-005 Intervention research: True experiments and quasi- experiments.
1 Women’s Studies. 2 First Wave Feminism Mary Wollstonecraft’s (1792) A Vindication of the Rights of Women is one of the first written works to be called.
The Miracle of Microfinance? Evidence from a Randomized Evaluation
The days ahead Monday-Wednesday –Training workshop on how to measure the actual reduction in HIV incidence that is caused by implementation of MC programs.
The Economic Effects of Social Security October 10-12, 2005.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Impact Evaluation in Education Introduction to Monitoring and Evaluation Andrew Jenkins 23/03/14.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Application 3: Estimating the Effect of Education on Earnings Methods of Economic Investigation Lecture 9 1.
MIRG Meeting 5: Impact of Microfinance Aruna Ranganathan.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Nigeria Impact Evaluation Community of Practice Abuja, Nigeria, April 2, 2014 Measuring Program Impacts Through Randomization David Evans (World Bank)
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
What is randomization and how does it solve the causality problem? 2.3.
Measuring Impact 1 Non-experimental methods 2 Experiments
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
S-005 Intervention research: True experiments and quasi- experiments.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
By Randall Munroe, xkcd.com Econometrics: The Search for Causal Relationships.
Randomized Evaluation: Start-to-finish Rachel Glennerster J-PAL.
Experimental Evaluations Methods of Economic Investigation Lecture 4.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
1 An introduction to Impact Evaluation (IE) for HIV/AIDS Programs March 12, 2009 Cape Town Léandre Bassolé ACTafrica, The World Bank.
Differences-in-Differences
Measuring Results and Impact Evaluation: From Promises into Evidence
Development Impact Evaluation in Finance and Private Sector
1 Causal Inference Counterfactuals False Counterfactuals
The Miracle of Microfinance? Evidence from a Randomized Evaluation
Presentation transcript:

Public Policy & Evidence: How to discriminate, interpret and communicate scientific research to better inform society. Rachel Glennerster Executive Director J-PAL Global

Press coverage of microcredit: 2000s

Many success stories…

We started to get some bad stories…

And some even worse…

Rigorous evidence: in between

Claim then reversal is common

The problem An NGO or politician promotes a program or policy There are good stories to tell:  you interview a woman who turned her life around after receiving credit Sometimes there is a correlation:  women who take out microfinance do better at business People assume this means the program or policy caused the change But “causal change” means change would not have happened in absence of the policy—hard to know Later studies find no effect, everyone is disillusioned with science and journalism

Different tools for different questions

Descriptive data describes conditions Examples: % of those in elite universities coming from poor families Good descriptive data are representative  press stories of skyrocketing food prices from Ebola (source: food seller near my hotel), but national survey showed no rise

Process data on implementation Did the textbooks arrive? Did subsidies go to the rich or to the poor? Example: take up rates of microcredit

Correlations: interesting not causal Women who give birth as teenagers are less likely to get good medical care and have more complications Women who give birth as teenagers also have less education, were doing less well in school before dropping out, come from poorer homes, etc etc Are they less likely to seek care because they give birth as teenagers or because they are less educated, poorer… This matters: if root problem is poverty then reducing teenage pregnancy may not reduce complications Correlations warn us of issues: don’t tell us how to fix them

Understanding causation is hard

2013 Intervention Start of the program Impact Year Causal impact requires knowing what would have happened without program

2013 Intervention Start of the program Impact Year Participants might be better off with the program Impact What would have happened

2013 Intervention Start of the program Impact Year Or they might have been worse off Impact

Example: Microcredit Comparing pre and post program assumes changes over time are due to the program  Many women are better off 5 years after taking out microcredit loans  But economy is growing, might have been better off over time even without microcredit What is wrong with studies comparing women in same village who took up microcredit vs those who did not?  How might those who took up microcredit be different? Power of statistical manipulation is limited.  Can compare women of similar age or education “observables”  No way to adjust for whatever made a woman take up microcredit

Spandana was expanding… Hyderabad Hyderabad, India

We randomly picked where Spandana would expand:

Microcredit: Randomization for evaluating impact Randomizing who has access to microcredit helps disentangle impact of microcredit from other factors:  Motivation  Entrepreneurial activity  Business opportunity growth Randomization creates two groups, treatment and comparison that (on average) are the same in all characteristics  Observable characteristics like income  Unobservable characteristics like motivation and entrepreneurial ability Can directly compare outcomes of those in treatment group and those in comparison group: difference is due to microcredit

Evidence: useful not transformative Seven randomized evaluations of microcredit in 7 countries Take up of microcredit was relatively modest amongst eligible borrowers In 2 of the studies there was an increase in the number of new businesses started Borrowers used the flexibility of credit: for example allowed them to purchase large durables and pay for them over time No increase in consumption No increase in women’s empowerment

Other methods for measuring impact Any study that seeks to measure causal impact must have a counterfactual:  Group that shows what would have happened without the program RCTs/experiments, comparison chosen at random Quasi-experimental methods require very specific assumptions to hold to be valid, hard to judge Examples of quasi-experimental methods  Regression discontinuity  Instrumental variables  Difference-in-difference JPAL is a resource on the assumptions needed for these different approaches to be valid

Telling a larger story with different kinds of evidence

Present bias and preventative health Descriptive evidence tells us there is a problem  The poor spend a lot of money on health but a lot of this is on ineffective acute interventions  There are many cost-effective prevention technologies that people don’t invest in eg good diet to hand washing  People bad at managing chronic disease Behavioral economic theories suggest a reason and solution:  Small inconsistencies about how we discount the future can cause big problems (place a lot of weight on today vs tomorrow but not tomorrow vs day after)  Small changes in price/convenience today will have big impacts RCTs show small changes in prices of preventative health have big impacts, and convenience very important (causal)

Journalists help people interpret claims

1. What type of claim is this: descriptive, process, correlation, or causal impact? 2. Does the type of evidence match the claim? Is the claim causal but the evidence only a correlation? 3. If the claim is descriptive, is the sample representative? If you only ask people who bought a product if they like it, its not telling you much. 4. If the claim is causal (ie about impact), what is the counterfactual—how do we know what would have happened otherwise? 5. Are there reasons to think that the people who participated in the program are different from those who did not participate? When reporting a claim always ask

1. Only use terms like: impact, led to, cause, or the result of when there is causal evidence—ie we know what would have happened in the absence of the program 2. Descriptive results are interesting and important, just be clear that they are only descriptive 3. Explain why a correlation is not necessarily causal: eg more of those on antidepressants commit suicide BUT those on antidepressants are depressed! 4. Academic studies can be written in complex language but the key ideas can often be explained in simple language, but to do so needs really understanding the results: 5. Ask authors how they would explain their results simply, run language by them, ask JPAL for help 6. Trust reliable journals but make sure to check your source (and do not hesitate to ask if you have doubts). Some tips for describing evidence

How can we promote use and understanding of evidence?