WHY LARGE-SCALE RANDOMIZED CONTROL TRIALS? David Myers Senior Vice President IES 2006 Research Conference David Myers Senior Vice President IES 2006 Research.

Slides:



Advertisements
Similar presentations
Responsiveness to Instruction North Carolina Problem Solving Model Problem Solving Model Session 1/4.
Advertisements

RTI: Questions and Answers June, Response to Intervention (RTI) What is it? a problem-solving systema problem-solving system a way to monitor progressa.
Teacher In-Service August, Abraham Lincoln.
Synthesizing the evidence on the relationship between education, health and social capital Dan Sherman, PhD American Institutes for Research 25 February,
Knowing if the RBF mechanism is working Incorporating Rigorous Impact Evaluation into your HRBF program Sebastian Martinez World Bank.
Using Data to Improve Schools One child at a time……….. The Educator’s Role Pamela M. Kastner National Board Certified Teacher.
Parents as Partners in Education
Pennsylvania’s Continuous Improvement Process. Understanding AYP How much do you know about AYP?
1 The Federal No Child Left Behind Act and the Financial Impact on Manchester Public Schools Fiscal Year
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
April 6, 2011 DRAFT Educator Evaluation Project. Teacher Education and Licensure DRAFT The ultimate goal of all educator evaluation should be… TO IMPROVE.
How Do We Know if a Charter School is Really Succeeding? – Various Approaches to Investigating School Effectiveness October 2012 Missouri Charter Public.
EVAAS EVALUATION EFFECTIVENESS …AND WHAT DOES IT SAY??? Brawley Middle School November 27, 2012.
VALUE – ADDED 101 Ken Bernacki and Denise Brewster.
RTI: Questions and Answers June, Response to Intervention (RTI) What is it? a problem-solving systema problem-solving system a way to monitor progressa.
Missing Data Issues in RCTs: What to Do When Data Are Missing? Analytic and Technical Support for Advancing Education Evaluations REL Directors Meeting.
Braiding Initiatives Steve Goodman, Michigan’s Integrated Behavior and Learning Initiative (MiBLSi) April 16, :00PM – 3:30PM
1 Implementation of Randomized Trials David Myers American Institutes for Research Washington, DC Prepared for IES/NCER Summer Research Training Institute,
1 Managing Threats to Randomization. Threat (1): Spillovers If people in the control group get treated, randomization is no more perfect Choose the appropriate.
Technology’s Edge: The Educational Benefits of Computer-Aided Instruction Lisa Barrow Federal Reserve Bank of Chicago Lisa Markman Princeton University.
Inclusive assessment and accountability systems: Issues and opportunities for students with disabilities in standards-based reform Rachel F. Quenemoen.
IES Grant Writing Workshop for Efficacy and Replication Projects
INSTRUCTIONAL LEADERSHIP FOR DIVERSE LEARNERS Susan Brody Hasazi Katharine S. Furney National Institute of Leadership, Disability, and Students Placed.
Agenda: Block Watch: Random Assignment, Outcomes, and indicators Issues in Impact and Random Assignment: Youth Transition Demonstration –Who is randomized?
Keystone Instructional Specialist. Keystone Exams Offered three times each year – winter, spring and summer. Offered in Algebra I, Biology and Literature.
Evaluation of Math-Science Partnership Projects (or how to find out if you’re really getting your money’s worth)
Common Questions What tests are students asked to take? What are students learning? How’s my school doing? Who makes decisions about Wyoming Education?
Creating Assessments with English Language Learners in Mind In this module we will examine: Who are English Language Learners (ELL) and how are they identified?
Tobacco Control Interventions – Design Trade-Offs K. S. (Steve) Brown Department of Statistics and Actuarial Science Health Behaviour Research Group University.
Inclusion or Mainstreaming Jenn Combest and Liz Raymer.
Overall Teacher Judgements
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Measuring Impact: Experiments
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
The Evaluation of Charter School Impacts June 30, 2010 Presentation at the 2010 IES Research Conference Philip Gleason ● Melissa Clark Christina Clark.
Targeted Assistance Programs: Requirements and Implementation Spring Title I Statewide Conference May 15, 2014.
What Was Learned from a Second Year of Implementation IES Research Conference Washington, DC June 8, 2009 William Corrin, Senior Research Associate MDRC.
Criteria for Assessing The Feasibility of RCTs. RCTs in Social Science: York September 2006 Today’s Headlines: “Drugs education is not working” “ having.
Response to Intervention: Improving Achievement for ALL Students Understanding the Response to Intervention Process: A Parent’s Guide Presented by: Dori.
1 Race to the Top Assessment Program General & Technical Assessment Discussion Jeffrey Nellhaus Deputy Commissioner January 20, 2010.
A Principled Approach to Accountability Assessments for Students with Disabilities CCSSO National Conference on Student Assessment Detroit, Michigan June.
Data for Student Success Using State Assessment Data to Identify School Improvement Goals Lani Seikaly Professional Development Coordinator Data for Student.
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
Educable Mental Retardation as a Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
RTI Response To Intervention. What is RTI ? Response to intervention is a multi – tier approach to the early identification and support of students with.
Evidence-based Education and the Culture of Special Education Chair: Jack States, Wing Institute Discussant: Teri Palmer, University of Oregon.
RtII Response to Instruction and Intervention paTTAN.
Getting Inside the “Black Box” – Capitalizing on Natural and Random Variation to Learn from the HPOG Impact Study Presenters: Alan Werner, co-Principal.
Elizabeth Spier, PhDJohannes Bos, PhD Principal ResearcherSenior Vice President FAST in Philadelphia SEPTEMBER 2015 Copyright © 2015 American Institutes.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
An Expanded Model of Evidence-based Practice in Special Education Randy Keyworth Jack States Ronnie Detrich Wing Institute.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Three ‘R’s for Evaluating the Memphis Striving Readers Project: Relationships, Real-World Challenges, and RCT Design Jill Feldman, RBS Director of Evaluation.
Overview of Student Learning Objectives (SLOs) for
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
Open Forum: Scaling Up and Sustaining Interventions Moderator: Carol O'Donnell, NCER
Forum on Evaluating Educator Effectiveness: Critical Considerations for Including Students with Disabilities Lynn Holdheide Vanderbilt University, National.
Vision: Every child in every district receives the instruction that they need and deserve…every day. Oregon Response to Intervention Vision: Every child.
Common Pitfalls in Randomized Evaluations Jenny C. Aker Tufts University.
Aim: Does the US need to reform the educational system? Do Now: Make a list of the best aspects of the education you receive and make a list of the worst.
IMPACT EVALUATION PBAF 526 Class 5, October 31, 2011.
Randomized Evaluation: Dos and Don’ts An example from Peru Tania Alfonso Training Director, IPA.
Measuring Results and Impact Evaluation: From Promises into Evidence
Duncanville ISD Curriculum Update
Comprehensive Planning
Implementation of Randomized Trials
SEC & State/District Initiatives
Impact Evaluation Designs for Male Circumcision
Sample Sizes for IE Power Calculations.
Presentation transcript:

WHY LARGE-SCALE RANDOMIZED CONTROL TRIALS? David Myers Senior Vice President IES 2006 Research Conference David Myers Senior Vice President IES 2006 Research Conference

Presentation Road Map Characteristics Benefits relative to small-scale RCTs Implementation and design challenges Characteristics Benefits relative to small-scale RCTs Implementation and design challenges

Characteristics Random assignment of schools/classrooms: –Multiple school districts (Reading Comprehension has about 10 districts; P4K has 27) –10 to 20 schools per “intervention arm” –200+ classrooms per “intervention arm” Random assignment of schools/classrooms: –Multiple school districts (Reading Comprehension has about 10 districts; P4K has 27) –10 to 20 schools per “intervention arm” –200+ classrooms per “intervention arm”

Characteristics (continued) Random assignment of students to treatment or control status with many schools/sites: –Upward Bound evaluation included a probability sample of 67 grantees and more than 3,000 students. –Charter school evaluation includes about 40 schools and 4,000 students. Random assignment of students to treatment or control status with many schools/sites: –Upward Bound evaluation included a probability sample of 67 grantees and more than 3,000 students. –Charter school evaluation includes about 40 schools and 4,000 students.

Characteristics (continued) Multiple years of data collection: –Baseline –One or more follow-up assessments Multiple interventions: –Reading Comp. and P4K: four each –Math Curriculum: three to four Multiple years of data collection: –Baseline –One or more follow-up assessments Multiple interventions: –Reading Comp. and P4K: four each –Math Curriculum: three to four

Benefits Impacts of school interventions, curricula, instruction, and professional development implemented in typical schools: –Often smaller impacts in large-scale studies (P4K) than in efficacy trials Impacts from a wider range of contexts than generally found in efficacy trials: –Students –Teachers –Schools –Greater opportunity for subgroup analyses Impacts of school interventions, curricula, instruction, and professional development implemented in typical schools: –Often smaller impacts in large-scale studies (P4K) than in efficacy trials Impacts from a wider range of contexts than generally found in efficacy trials: –Students –Teachers –Schools –Greater opportunity for subgroup analyses

What Questions Can an RCT Address? Intention to treat – yes Treatment on the treated – yes, with assumptions Ingredients and dosage – no design answers particularly well Intention to treat – yes Treatment on the treated – yes, with assumptions Ingredients and dosage – no design answers particularly well

Intention to Treat Impact of the opportunity to participate for all schools/teachers/students Simply, the difference in the means for the treated group and the control group Control group represents what we expect for treated group in absence of intervention Most policy relevant Subgroup analysis shows who benefits most Impact of the opportunity to participate for all schools/teachers/students Simply, the difference in the means for the treated group and the control group Control group represents what we expect for treated group in absence of intervention Most policy relevant Subgroup analysis shows who benefits most

Treatment on the Treated More assumptions Impact of intervention on those who actually receive treatment Adjusts for those who did not participate and control who did Less policy relevant because it does not show the impact of an implemented policy – TOT separates the impact on participation from the effect of the treatment Subgroup analysis shows who benefits most More assumptions Impact of intervention on those who actually receive treatment Adjusts for those who did not participate and control who did Less policy relevant because it does not show the impact of an implemented policy – TOT separates the impact on participation from the effect of the treatment Subgroup analysis shows who benefits most

Ingredients/Dosage Many assumptions, some unrealistic Like other designs, cannot address these questions with the same level of rigor Random assignment to different “intervention arms” of ingredients/doses needed for defensible impact estimates Many assumptions, some unrealistic Like other designs, cannot address these questions with the same level of rigor Random assignment to different “intervention arms” of ingredients/doses needed for defensible impact estimates

Study Implementation Challenges Objections involve interventions, not design, so negotiation teams must have substantive expertise More innovation in schools in recent years – less inclined to adopt yet another approach Districts want consistent instruction across schools Alignment of curriculum with state testing NCLB waivers – will schools and districts be held accountable (AYP) if the intervention fails? Objections involve interventions, not design, so negotiation teams must have substantive expertise More innovation in schools in recent years – less inclined to adopt yet another approach Districts want consistent instruction across schools Alignment of curriculum with state testing NCLB waivers – will schools and districts be held accountable (AYP) if the intervention fails?

Study Implementation Challenges (cont.) How much “intervention from the evaluator” is OK when it comes to fidelity? –Let policy and practice dictate decisions – what would developer or school do in the absence of the evaluation? Maintain normal school, program, and intervention operations – make the design accommodate the program When oversubscription is needed, programs and schools often overestimate student interest. How much “intervention from the evaluator” is OK when it comes to fidelity? –Let policy and practice dictate decisions – what would developer or school do in the absence of the evaluation? Maintain normal school, program, and intervention operations – make the design accommodate the program When oversubscription is needed, programs and schools often overestimate student interest.

Generic Design Challenges External validity – random effects design requires larger samples for same precision as fixed effects – demos vs. “national evaluations” Spillover or contamination from random assignment of schools to interventions Effect of giving some students treatment on the outcomes for the control group (P4K) Identifying the counterfactual External validity – random effects design requires larger samples for same precision as fixed effects – demos vs. “national evaluations” Spillover or contamination from random assignment of schools to interventions Effect of giving some students treatment on the outcomes for the control group (P4K) Identifying the counterfactual

Design Challenges: Case Study Charter schools evaluation: –Research question concerns the impact of attending a charter school  How would a child attending a charter school have performed in the absence of that school? –Design is random assignment to charter/regular public schools (lottery) –All designs are limited here:  A nonrandom sample of students left the public school, which may change what is taught there  Different than the thought experiment of asking what would happen with no charter schools Charter schools evaluation: –Research question concerns the impact of attending a charter school  How would a child attending a charter school have performed in the absence of that school? –Design is random assignment to charter/regular public schools (lottery) –All designs are limited here:  A nonrandom sample of students left the public school, which may change what is taught there  Different than the thought experiment of asking what would happen with no charter schools

Another Case Study: P4K What is the impact of being assigned to a pullout program? –How would students perform in the absence of the pullout program? Pullout program: –RA of schools to one of four interventions –Within schools, RA of eligible students to the pullout program or the regular classroom Change in the composition of the regular classroom and perhaps, change in curriculum and instruction when some struggling readers are removed What is the impact of being assigned to a pullout program? –How would students perform in the absence of the pullout program? Pullout program: –RA of schools to one of four interventions –Within schools, RA of eligible students to the pullout program or the regular classroom Change in the composition of the regular classroom and perhaps, change in curriculum and instruction when some struggling readers are removed