Impact Evaluation “Randomized Evaluations” Jim Berry Asst. Professor of Economics Cornell University.

Slides:



Advertisements
Similar presentations
Questions What is the relationship between ‘research designs’ and ‘research strategies’? Which method of experiments, within subjects or between subjects.
Advertisements

Experimental Design True Experimental Designs n Random assignment n Two comparison groups n Controls threats to internal validity n Strongest evidence.
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
Validity (cont.)/Control RMS – October 7. Validity Experimental validity – the soundness of the experimental design – Not the same as measurement validity.
Types of Experimental Designs Non-experimental –Post-test only –Pre-post test Experimental –Pre & post-test with control group –Pre & multiple post-test.
What could go wrong? Deon Filmer Development Research Group, The World Bank Evidence-Based Decision-Making in Education Workshop Africa Program for Education.
Designing Influential Evaluations Session 5 Quality of Evidence Uganda Evaluation Week - Pre-Conference Workshop 19 th and 20 th May 2014.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Research Design and Validity Threats
Childhood Obesity Scenario: Quasi- Experiments and Natural Experiments Versus RCTs Steven Gortmaker, Ph.D. Harvard School of Public Health /Harvard Prevention.
1 Managing Threats to Randomization. Threat (1): Spillovers If people in the control group get treated, randomization is no more perfect Choose the appropriate.
Validity, Sampling & Experimental Control Psych 231: Research Methods in Psychology.
Who are the participants? Creating a Quality Sample 47:269: Research Methods I Dr. Leonard March 22, 2010.
Non-Experimental designs: Developmental designs & Small-N designs
Sampling and Experimental Control Goals of clinical research is to make generalizations beyond the individual studied to others with similar conditions.
Clinical Trials Hanyan Yang
TOOLS OF POSITIVE ANALYSIS
Agenda: Block Watch: Random Assignment, Outcomes, and indicators Issues in Impact and Random Assignment: Youth Transition Demonstration –Who is randomized?
Experiments and Observational Studies.  A study at a high school in California compared academic performance of music students with that of non-music.
Chapter 8 Experimental Research
1 Randomization in Practice. Unit of randomization Randomizing at the individual level Randomizing at the group level –School –Community / village –Health.
I want to test a wound treatment or educational program in my clinical setting with patient groups that are convenient or that already exist, How do I.
Research Design for Quantitative Studies
Group Discussion Explain the difference between assignment bias and selection bias. Which one is a threat to internal validity and which is a threat to.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Sociology 3322a. “…the systematic assessment of the operation and/or outcomes of a program or policy, compared to a set of explicit or implicit standards.
Randomized Control Trials for Agriculture Pace Phillips, Innovations for Poverty Action
Chapter 13 Observational Studies & Experimental Design.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 13 Experiments and Observational Studies.
TRANSLATING RESEARCH INTO ACTION What is Randomized Evaluation? Why Randomize? J-PAL South Asia, April 29, 2011.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
MARKETING RESEARCH CHAPTERS
ECON ECON Health Economic Policy Lab Kem P. Krueger, Pharm.D., Ph.D. Anne Alexander, M.S., Ph.D. University of Wyoming.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)
POLS 7170X Master’s Seminar Program/Policy Evaluation Class 5 Brooklyn College – CUNY Shang E. Ha.
Copyright ©2008 by Pearson Education, Inc. Pearson Prentice Hall Upper Saddle River, NJ Foundations of Nursing Research, 5e By Rose Marie Nieswiadomy.
How do the Freshmen Feel? A survey conducted by Caitlin Park and Eli Harris.
LAURA RALSTON, ECONOMIST, CCSD FINDING TRUE PROGRAM IMPACTS THROUGH RANDOMIZATION.
Single-Subject Experimental Research
Potential Errors In Epidemiologic Studies Bias Dr. Sherine Shawky III.
Ch. 2 Tools of Positive Economics. Theoretical Tools of Public Finance theoretical tools The set of tools designed to understand the mechanics behind.
CHAPTER 12 Descriptive, Program Evaluation, and Advanced Methods.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Understanding Sampling
Applying impact evaluation tools A hypothetical fertilizer project.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Measuring Impact 1 Non-experimental methods 2 Experiments
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
Experimental & Quasi-Experimental Designs Dr. Guerette.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Chapter Eight: Quantitative Methods
October 16, 2002Experimental Design1 Principles of Experimental Design October 16, 2002 Mark Conaway.
Design of Clinical Research Studies ASAP Session by: Robert McCarter, ScD Dir. Biostatistics and Informatics, CNMC
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Unit 4: Gathering Data LESSON 4-4 – EXPERIMENTAL STUDIES ESSENTIAL QUESTION: WHAT ARE GOOD WAYS AND POOR WAYS TO EXPERIMENT?
Uses of Diagnostic Tests Screen (mammography for breast cancer) Diagnose (electrocardiogram for acute myocardial infarction) Grade (stage of cancer) Monitor.
Impact Evaluation Methods Regression Discontinuity Design and Difference in Differences Slides by Paul J. Gertler & Sebastian Martinez.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Experimental Design-Chapter 8
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
2 independent Groups Graziano & Raulin (1997).
Quantitative Research
Impact Evaluation Methods
Presenter: Kate Bell, MA PIP Reviewer
Presentation transcript:

Impact Evaluation “Randomized Evaluations” Jim Berry Asst. Professor of Economics Cornell University

Randomized Evaluations Also known as: Random Assignment Studies Randomized Field Trials Social Experiments Randomized Controlled Trials (RCTs) Randomized Controlled Experiments 2

Start with simple case: Take a sample of program applicants Randomly assign them to either: Treatment Group – is offered treatment Control Group - not allowed to receive treatment (during the evaluation period) What does the term random mean? 3

Key advantage of experiments Because members of the groups (treatment and control) do not differ systematically at the outset of the experiment, any difference that subsequently arises between them can be attributed to the treatment rather than to other factors. 4

Relative to results from non-experimental studies, results from experiments are: Less subject to methodological debates Easier to convey More likely to be convincing to program funders and/or policymakers

 Despite great methodological advantage of experiments, they are also potentially subject to threats to their validity. For example,  Internal Validity (e.g. survey non-response, no-shows, crossovers, duration bias, etc.)  External Validity (e.g. are the results generalizable to other populations?)  It is important to realize that some of these threats also affect the validity of non- experimental studies 7

Design the study carefully Randomly assign people to treatment or control Collect baseline data Verify that assignment looks random Monitor process so that integrity of experiment is not compromised 8

Collect follow-up data for both the treatment and control groups in identical ways. Estimate program impacts by comparing mean outcomes of treatment group vs. mean outcomes of control group. Assess whether program impacts are statistically significant and practically significant. 9

Basic setup of a randomized evaluation Target PopulationPotential participantsEvaluation Sample Treatment Group ParticipantsNo-Shows Control group

What I haven’t told you: Standard 4 balsakhis were randomly assigned to schools: half got the balsakhi, half didn’t Strategy: compare low-performing students in balsakhi standards with low-performing students in non-balsakhi standards How might this help? Answers the question: What is the impact of a balsakhi on low-performing children?

When assignment is random, standards with a balsakhi should provide an unbiased control group Let’s repeat the exercise comparing low- performing children in balsakhi schools to low- performing children in no-balsakhi schools

Average post-test score for low-scoring students in balsakhi grades Average post-test score for low-scoring students in non- balsakhi grades Difference10.60

Another estimate of impact: having a balsakhi in your grade vs. not having a balsakhi in your grade, whether or not you learned from the balsakhi Mean Score for Children with a balsakhi Mean score for children without a balsakhi Difference6.45

Summary table 15 Source: Arceneaux, Gerber, and Green (2004) MethodEstimate 1 – Pre-Post – Simple Difference – Difference-in-Differences Regression2.3 5 – Randomized Simple Difference 10.60

If properly designed and conducted, social experiments provide the most credible assessment of the impact of a program Results from social experiments are easy to understand and much less subject to methodological quibbles Credibility + Ease of understanding =>More likely to convince policymakers and funders of effectiveness (or lack thereof) of program 16

However, these advantages are present only if social experiments are well designed and conducted properly Must assess validity of experiments in same way we assess validity of any other study Must be aware of limitations of experiments 17