Opportunities and Challenges in a Multi-Site Regression Discontinuity Design Stephen W. Raudenbush University of Chicago Presentation at the MultiLevel.

Slides:



Advertisements
Similar presentations
Multilevel Interventions: Study Design and Analysis Issues Paul D. Cleary, Ph.D. Yale School of Public Health Cary Gross, M.D. Yale School of Medicine.
Advertisements

Location of Uganda Proposed Impact Evaluation for Uganda.
#ieGovern Impact Evaluation Workshop Istanbul, Turkey January 27-30, 2015 Measuring Impact 1 Non-experimental methods 2 Experiments Vincenzo Di Maro Development.
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
The counterfactual logic for public policy evaluation Alberto Martini hard at first, natural later 1.
Regression Discontinuity Design William Shadish University of California, Merced.
Regression Discontinuity Design Thanks to Sandi Cleveland and Marc Shure (class of 2011) for some of these slides.
Conditions to Support Successful Teaching Challenging Coursework: School Climate and Organization Elaine Allensworth April 10, 2012.
WHY LARGE-SCALE RANDOMIZED CONTROL TRIALS? David Myers Senior Vice President IES 2006 Research Conference David Myers Senior Vice President IES 2006 Research.
Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Analysis of Clustered and Longitudinal Data
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Including a detailed description of the Colorado Growth Model 1.
Using Student Surveys: Implications for State and District Policy May 1, #aypfevents.
Evaluating a Literacy Curriculum for Adolescents: Results from Three Sites of the First Year of Striving Readers Eastern Evaluation Research Society Conference.
Causal Inference for Time-varying Instructional Treatments Stephen W. Raudenbush University of Chicago Joint Work with Guanglei Hong The research reported.
College Preparatory Curriculum for All Lessons Learned from Chicago Elaine Allensworth with TakakoNomi and Nicholas Montgomery at the Consortium on Chicago.
Achieved Relative Intervention Strength: Models and Methods Chris S. Hulleman David S. Cordray Presentation for the SREE Research Conference Washington,
Striving to Link Teacher and Student Outcomes: Results from an Analysis of Whole-school Interventions Kelly Feighan, Elena Kirtcheva, and Eric Kucharik.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Causal Inference and Adequate Yearly Progress Derek Briggs University of Colorado at Boulder National Center for Research on Evaluation, Standards, and.
Centre for Market and Public Organisation Using difference-in-difference methods to evaluate the effect of policy reform on fertility: The Working Families.
HISTORY OF ANEMIC PROGRESS OF ENGLISH LANGUAGE LEARNERS ON STATE ACCOUNTABILITY TESTS Manuel Barrera, PhD Associate Professor, Metropolitan State University;
© Michael Lechner, 2006, p. 1 (Non-bayesian) Discussion (translation) of Principal Stratification for Causal Inference with Extended Partial Complience.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
Problems with the Design and Implementation of Randomized Experiments By Larry V. Hedges Northwestern University Presented at the 2009 IES Research Conference.
ANNUAL LECTURE Heterogeneous Agents, Social Interactions and Causal Inference This talk is based on “Heterogeneous Agents, Social Interactions, and Causal.
Replacement Cases Framework overview Thresholds for inference and % bias to invalidate The counterfactual paradigm Internal validity example: kindergarten.
SUTVA, Assignment Mechanism STA 320 Design and Analysis of Causal Studies Dr. Kari Lock Morgan and Dr. Fan Li Department of Statistical Science Duke University.
Private involvement in education: Measuring Impacts Felipe Barrera-Osorio HDN Education Public-Private Partnerships in Education, Washington, DC, March.
Assumptions of value-added models for estimating school effects sean f reardon stephen w raudenbush april, 2008.
Introduction to Multilevel Modeling Stephen R. Porter Associate Professor Dept. of Educational Leadership and Policy Studies Iowa State University Lagomarcino.
© CCSR The CCSR 5 Essentials Survey Helping Schools Organize for Improvement Elaine Allensworth The University of Chicago Consortium on Chicago School.
Optimal Design for Longitudinal and Multilevel Research Jessaca Spybrook July 10, 2008 *Joint work with Steve Raudenbush and Andres Martinez.
ECON 3039 Labor Economics By Elliott Fan Economics, NTU Elliott Fan: Labor 2015 Fall Lecture 21.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
What is randomization and how does it solve the causality problem? 2.3.
Talkin’ Hard Luck “I’ve been cussed and discussed, Talked to and talked about, Lied to and lied about, Held up and hung up And doggone nigh murdered, And.
Studying the Mean and Variation in the Effect of Program Participation in Multi-site Trials The research reported here was supported by a grant from the.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
The Power of Comparison in Learning & Instruction Learning Outcomes Supported by Different Types of Comparisons Dr. Jon R. Star, Harvard University Dr.
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
Impact Evaluation for Evidence-Based Policy Making Arianna Legovini Lead Specialist Africa Impact Evaluation Initiative.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Randomized Control Trials: What, Why, How, When, and Where Mark L. Davison MESI Conference March 11, 2016 Department of Educational Psychology.
Cross-Country Workshop for Impact Evaluations in Agriculture and Community Driven Development Addis Ababa, April 13-16, Causal Inference Nandini.
PRAGMATIC Study Designs: Elderly Cancer Trials
David M. Murray, Ph.D. Associate Director for Prevention Director, Office of Disease Prevention Multilevel Intervention Research Methodology September.
Classroom Network Technology as a Support for Systemic Mathematics Reform: Examining the Effects of Texas Instruments’ MathForward Program on Student Achievement.
Measuring Results and Impact Evaluation: From Promises into Evidence
Stephen W. Raudenbush University of Chicago December 11, 2006
Econometrics ITFD Week 8.
An introduction to Impact Evaluation
Difference-in-Differences
Explanation of slide: Logos, to show while the audience arrive.
Impact evaluation: The quantitative methods with applications
EMPIRICAL STUDY AND FORECASTING (II)
Evaluating Impacts: An Overview of Quantitative Methods
Explanation of slide: Logos, to show while the audience arrive.
Sampling for Impact Evaluation -theory and application-
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
CCSSO National Conference on Student Assessment June 21, 2010
Counterfactual models Time dependent confounding
Chapter 3 Hernán & Robins Observational Studies
Presentation transcript:

Opportunities and Challenges in a Multi-Site Regression Discontinuity Design Stephen W. Raudenbush University of Chicago Presentation at the MultiLevel Theory and Research Conference The Pennsylvania State University University Park, PA, May 17, 2015 The research reported here was supported by a grant from the WT Grant Foundation entitled “Learning from Variation In Program Effects: Methods, Tools, and Insights from Multi-site Trials.”

Outline Counter-factual account of causation The “drug-trial paradigm” for causal inference An alternative paradigm for social interventions Heterogeneous agents Social interactions among participants Curriular reform in chicago Conventional RDD Incorporating Agents and Social Interactions Identification: School-specific IV Conclusions

Counter-factual account of Causation In statistics (Neyman, Rubin, Rosenbaum) In economics (Haavelmo, Roy, Heckman)

Drug trial paradigm for causation Y (1) : Outcome if the patient receives Z = 1 (the “new drug”) Y (0) : Outcome if the patient receives Z = 0 (the “standard treatment”) Y (1) – Y (0) : Patient-specific causal effect E (Y (1) – Y (0) ) =  : Average causal effect

Stable Unit Treatment Value Assumption (Rubin, 1986) Each patient has two potential outcomes Implies – Only one “version” of each treatment – No “interference between units” Implies the doctor and the other patients have no effect on the potential outcomes

Formally…

Failure of SUTVA in Education Teachers enact instruction in classrooms –Multiple “versions of the treatment” Treatment assignment of one’s peers affects one’s own potential outcomes –EG Grade Retention –Hong and Raudenbush, Educational Evaluation and Policy Analysis, 2005 –Hong and Raudenbush, Journal of the American Statistical Association, 2006

Group-Randomized Trials Potential outcome Thus, each child has only two potential outcomes – if we have “intact classrooms” – if we have “no interference between classrooms”

Limitations of cluster randomized trial Mechanisms operate within clusters * Example: 4Rs teachers vary in response classroom interactions spill over We may have interference between clusters * Example: community policing

Alternative Paradigm for social interventions Treatment setting (Hong, 2004): A unique local environment for each treatment composed of * a set of agents who may implement an intervention and * a set of participants who may receive it Each participant possesses a single potential outcome within each possible treatment setting Causal effects are comparisons between these potential outcomes

Example: Community Policing (Verbitsky-Shavitz and Raudenbush, 2012) Let Z j =1 if Neighborhood j gets community policing Let Z j =0 if not Under SUTVA

“All or none”

Do it only in high-crime areas: effect on low-crime areas 1, HC 0, LC 1, HC 0, HC 0, LC 0, HC 0, LC

Results Having community policing was especially good if your surrounding neighbors had it Not having community policing was especially bad if your neighbors had it *** So targetting only high crime areas may fail***

Application: Double-dose Algebra Nomi and Raudenbush (2015) Requires 9 th -graders to take Double-dose Algebra if they scored below 50 percentile on 8 th -grade math test 12,000 students in 60 Chicago high schools

Double-dose Algebra enrollment rate by math percentile scores (city wide) Enrollment Rates ITBS percentile scores

Conventional Mediation Model (T, M,Y model) Cut off (T) Double-Dose Algebra (M) Algebra Learning (Y) γ = average effect of cutoff on taking Double Dose δ=average effect of taking Double Dose on Y for compliers (“CACE” effect) Assume no direct effect of T on Y (exclusion restriction) β= γδ (“ITT” effect) So δ= β/ γ Nomi, T., & Allensworth, E. (2009) γ δ

Conventional Model is Founded on SUTVA

Results of Conventional Analysis Large average impact of Cut on taking DD (ITT effect on M) Modest average impact of Cut on Y (ITT effect on Y) Modest CACE (Average Impact of M on compliers)

ITT effect on Y

But the policy changed classroom composition!!

Classroom average skill levels by math percentile scores Pre-policy ( and cohorts) Post-policy ( and cohorts)

Implementation varied across schools in--- Complying with the policy Inducing classroom segregation

Exclusion Restriction Revised T-M-C-Y model Cut off (T) Double-Dose Algebra (M) Algebra score (Y) Classroom Peer skill (C)

Identification Problem We have one equation, two unknowns: Strategy is school-specific

A simple two-level model At level 1 At level 2

Derivation of assumptions using potential outcomes

ParameterEstimateSE ITT impact on M ITT impact on C ITT impact on Y CACE of M CACE of C

5. Conclusions on DD The reform Increased instructional time Changed class composition Median skill kids Gained a lot if not tracked into low-skill classes Gained little if they were

Conclusions on Causal Inference Conventional causal paradigm: * a single potential outcome per participant under each treatment Alternative paradigm: * a single potential outcome per participant in each treatment setting RDD as a means-tested program Potentially large policy implications of causal paradigm