Challenges arising from the analysis of randomized trials in education

Slides:



Advertisements
Similar presentations
Designing an impact evaluation: Randomization, statistical power, and some more fun…
Advertisements

The Education Endowment Foundation (EEF) is an independent grant-making charity dedicated to breaking the link between family income and educational achievement,
USING EVIDENCE TO INFORM YOUR LEADERSHIP APPROACH AND SUPPORT SCHOOL IMPROVEMENT ROB CARPENTER 26 TH SEPTEMBER 2013
Dorset Leadership Conference, 2013 Using evidence to inform your leadership approach and support school improvement James Richardson 5 th November 2013.
∂ What works…and who listens? Encouraging the experimental evidence base in education and the social sciences RCTs in the Social Sciences 9 th Annual Conference.
Professor Steve Higgins School of Education, Durham University
Exploring Research-Led Approaches to Increasing Pupil Learning Steve Higgins School of Education, Durham University Addressing.
Building Evidence in Education: Workshop for EEF evaluators 2 nd June: York 6 th June: London
Effective use of the Pupil Premium to close the attainment gap James Richardson Senior Analyst, Education Endowment Foundation 27 th June 2014
Adapting Designs Professor David Torgerson University of York Professor Carole Torgerson Durham University.
Using evidence to raise the attainment of children facing disadvantage James Richardson Senior Analyst, Education Endowment Foundation 1 st April 2014.
How can evidence contribute to closing the attainment gap? James Richardson & Jonathan Sharples Education Endowment Foundation 16 th March 2015
Campbell Collaboration Colloquium 2012 Copenhagen, Denmark The effectiveness of volunteer tutoring programmes Dr Sarah Miller Centre.
Methods and Approaches to investigate the UK Education System Sandra McNally, University of Surrey and Centre for Economic Performance, London School of.
Making all research results publically available: the cry of systematic reviewers.
School of Education Archive Analysis: On EEF Trials Adetayo Kasim, ZhiMin Xiao, and Steve Higgins.
Using research to get the best value from the Pupil Premium Steve Higgins, School of Education, Durham National.
Research evidence and effective use of the Pupil Premium Professor Steve Higgins, School of Education, Durham
Addressing educational disadvantage, sharing evidence, finding out what works Camilla Nevill Evaluation Manager.
1 Effective use of the pupil premium to raise achievement Lessons from research Robbie Coleman 18 th June 2015.
Planning high quality, evidence based provision to meet the needs and achieve the outcomes How do you know what works?
Building Evidence in Education: Conference for EEF Evaluators 11 th July: Theory 12 th July: Practice
The Pupil Premium: Using Evidence to Narrow the Gap Robbie Coleman 7 th July 2014
The Campbell Collaborationwww.campbellcollaboration.org C2 Training: May 9 – 10, 2011 Introduction to meta-analysis.
The Education Endowment Foundation Kevan Collins 26th September 2013
Using Evidence to Narrow the Gaps. What is the Education Endowment Foundation? In 2011 the Education Endowment Foundation was set up by Sutton Trust as.
Retain H o Refute hypothesis and model MODELS Explanations or Theories OBSERVATIONS Pattern in Space or Time HYPOTHESIS Predictions based on model NULL.
Impact of two teacher training programmes on pupils’ development of literacy and numeracy ability: a randomised trial Jack Worth National Foundation for.
Developing teaching as an evidence informed profession UCET Annual Conference Kevan Collins - Chief Executive
Replication in Prevention Science Valentine, et al.
The rise of RCTs in education: statistical and practical challenges Kevan Collins.
A Quick Guide to Pupil Premium Spending. Interesting Data Based on 2013 data the gap between FSM and non FSM students gets wider as students get older.
Approaches to quantitative data analysis Lara Traeger, PhD Methods in Supportive Oncology Research.
The Educational Endowment Foundation (EEF) Application for funding to raise the attainment of FSM pupils Breakfast Briefing 5 th October 2011 Catherine.
Planning high quality, evidence based provision to meet the needs and achieve the outcomes How do you know what works?
Planning Key Stage 3 National Literacy Trust
Evaluation in Education: 'new' approaches, different perspectives, design challenges Camilla Nevill Head of Evaluation, Education Endowment Foundation.
2016 Primary Assessment Update 27th September 2016
How to give research results back to teachers?
Non-Housing Outcomes of Accessible and Affordable Housing
SLP Training Day 3 30th September 2016
Professor Steve Higgins, School of Education, Durham University
Dissemination and scale-up is the key to the EEF’s impact.
Paula Clarke, Charles Hulme, Emma Truelove, Maggie Snowling
The English RCT of ‘Families and Schools Together’
Benefits and Pitfalls of Systematic Reviews and Meta-Analyses
Reducing bias in randomised controlled trials involving therapists:
Supplementary Table 1. PRISMA checklist
EEF Archive analysis overview
Lecture 4: Meta-analysis
Evidence in Action: Using Research to Narrow the Gap Eleanor Stringer
Next steps for raising attainment: leadership, school autonomy and improving standards in the classroom Allen Thurston, Professor of Education & Liam O’Hare,
Challenges of statistical analysis in surgical trials
STROBE Statement revision
Statistical Analysis Plan review
Early Years – early language, social mobility and the home learning environment 15 March 2018.
The Education Endowment Foundation
Moving towards evidence-based policy
Systematic review and meta-analysis
Gerald Dyer, Jr., MPH October 20, 2016
Analysing educational trials: the Education Endowment Foundation Archive Steve Higgins, Adetayo Kasim, ZhiMin Xiao, with Nasima Akhter, Ewoud De Troyer,
Research & scholarship
The impact of small-group EBP education programme: barriers and facilitators for EBP allied health champions to share learning with peers.
EAST GRADE course 2019 Introduction to Meta-Analysis
Methods and Approaches to investigate the UK Education System
Applying to the EEF for funding: What are we looking for
Benchmarking and Collaboration
Finance – making the best of your resources budget planning, benchmarking, collaboration & seeking best value Welcome.
CCSSO National Conference on Student Assessment June 21, 2010
The EEF approach Sir Kevan Collins 25th March 2019.
Presentation transcript:

Challenges arising from the analysis of randomized trials in education Professor Steve Higgins s.e.higgins@durham.ac.uk School of Education Durham University Randomised Controlled Trials in the Social Sciences 7 - 9 September 2016, University of York

Overview The work of the Education Endowment Foundation The EEF Archive Analysis project (with Adetayo Kasim and ZhiMin Xiao) Trials findings and the Toolkit Some implications for research, policy and practice

The work of the Education Endowment Foundation Established in 2011 with £125M endowment (aim £200M total) Independent grant-making charity dedicated to breaking the link between family income and educational achievement Identify, test and scale successful approaches and interventions Meta-analytic evidence database to identify promise (from the ‘Teaching and Learning Toolkit’) – “What Works Centre for improving education outcomes for school-aged children” Projects commissioned from a range of organisations Independently evaluated by a team from a ‘panel’ of 25 Results fed back into Toolkit (with other similar project findings and new meta-analyses)

Projects and campaigns Pilot Efficacy Effectiveness Campaigns Making Best Use of Teaching Assistants (£5M) North East Literacy Campaign (with Northern Rock Foundation - £10M)

EEF since 2011 60 independent evaluation reports published 64% £75.4m funding awarded to date 64% of school leaders say they have used the toolkit 7,500 schools participating in projects 750,000 pupils currently involved in EEF projects 60 independent evaluation reports published 106 RCTs 127 projects funded to date

The EEF Archive Analysis project Independent evaluators submit project data to FFT (given appropriate permissions) Data re-matched with National Pupil Database (NPD) data and updated as new data becomes available Released to archive team at Durham twice a year Aims to undertake methodological exploration of trials data provide comparative analyses of impact across projects analyse long term/ follow-up impact as NPD data becomes available develop an R package for educational trials analysis

Principles and caveats Principles Exploratory analysis to inform EEF’s evaluation strategy Help to explain variation in trial impact Independent evaluator’s peer-reviewed published estimate is always the official EEF finding for impact on attainment Caveats Can’t always match N (4/32 projects greater than 10% difference) NPD data removed and re-matched Some raw scores transformed Gain score distribution different from post-test

R package - eefAnalytics To support analysis of randomised trials in education (individual, cluster and multi-site) difference-in-means ordinary least squares multi-level models (frequentist & Bayesian) Permutation p-values, bootstrapped confidence intervals Complier Average Causal Effect (CACE) Cumulative quantile analysis eefAnalytics: in development – available on CRAN in November – available to try out – contact us!

Initial findings 17 EEF projects Four analytic models Findings Results converge in larger, unproblematic trials Point estimates and estimates of precision vary when trials are problematic (e.g. testing issues, randomisation), when design and analysis not matched; with different co-variates added; when outcomes are different or transformed (e.g. z-scores) Results tend to diverge if ICC ≥ 0.2 and there are few clusters/schools MLM total variance ‘most conservative’ model (wider CI) Bayesian estimates identical but more precise (narrower CI)

Convergence Pre-test imbalance

Clustering Point estimates similar, CIs vary

Divergence Pre-test imbalance Post-ANCOVA models aim to correct for this Evaluator gain estimate higher

Analytic heterogeneity Stem Leaf -0.1 -0.0 0.0 0 0 1 1 1 1 2 3 3 4 4 4 4 5 5 5 6 6 6 7 7 7 7 8 0 0 0 0 0 1 1 1 1 1 2 2 2 3 4 4 4 4 5 5 5 5 5 6 7 7 7 8 8 8 0.1 0 1 1 5 7 7 7 8 0.2 4 Stem and leaf plot of differences between evaluator and archive MLM (total variance) models 32 projects 64 outcomes Majority 0.05 or less

Archive Analysis development R package release - November 2016 Post/gain paper (revisiting Lord’s paradox) Local Influence Index (a binomial index for evaluating intervention benefit) CACE (to help interpret ITT analysis) Follow-up data in NPD Reflections With ITT approach, sample sizes predicated on minimum necessary to detect a probably overestimated effect size, and MLM total variance are we setting ourselves up for disappointment?

Sutton Trust/EEF Teaching & Learning Toolkit Best ‘buys’ on average from research Key messages for Pupil Premium spending in schools Currently used by over 60% of school leaders http://educationendowmentfoundation.org.uk/toolkit

Toolkit overview and aims Cost effectiveness estimates of a range of educational approaches Based on average effects from meta-analyses and cost estimates of additional outlay to put in place Evidence robustness estimates as ‘padlocks’ To inform professional decision-making about school spending To create a framework for evidence-use To provide a structure to improve evidence utility

Aggregating inferences

Inferences from findings across meta-analyses Requires assumption of variation, bias and inaccuracy randomly distributed across included studies Probably unwarranted, but the best we’ve got Starting point to improve precision and predictability Better than saying nothing?

Toolkit as a predictor? Toolkit theme EEF Projects Effect Size Digital technology 0.28 Phonics 0.35 Accelerated Reader 0.24 Butterfly Phonics 0.43 Improving Lit and Num 0.20 Fresh start Units of Sound -0.08 Catch up literacy 0.12 Vocabulary Enrichment 0.07 Rapid Phonics -0.05 Meta-cogn & SRL 0.62 Teaching assistants 0.08 Improving writing quality 0.74 Catch up numeracy 0.21 Changing Mindsets 0.18 Talk for Literacy P4C One to one 0.40 Summer school Perry Beeches 0.36 Future Foundations (Eng) 0.17 Switch on Reading Future Foundations (maths) 0.00 Summer Active Reading 0.14 Peer tutoring Shared Maths 0.02 Paired reading -0.02 *As we can see overall the impact of the specific programmes come in agreement with the Toolkit topic!

So, “what works” or “what’s worked”? Internal validity necessary for external – did it actually work there? Causal ‘black-box’ makes replicability challenging Defining ‘approaches’ or ‘interventions’ – unit of description Problematic ‘populations’ – what inference for whom? Importance of knowing what hasn’t worked (on average) In education a null result (i.e. may not be different from 0) = as good as counterfactual Mean or range – “on average” or better estimates of probability? Generalisability or predictability? Small-scale research findings may optimise rather than typify ‘in the wild’ application or scale-up

A distributional view Visualised distribution of Toolkit effects Distribution of effects in Kluger & de Nisi (1996) from Dylan Wiliam (https://twitter.com/dylanwiliam/status/608610040086409216 ) Visualised distribution of Toolkit effects

Current Toolkit developments Formalising methodology (translating/simplifying existing models) Cochrane/ Campbell/ EPPI PRISMA for reviews CONSORT for trials GRADE Guidelines for evidence New comparable and updatable meta-analyses for each strand Identifying factors affecting current effect size estimates Design (sample size, randomisation, clustering) Measurement issues (outcome complexity, outcome alignment) Intervention (duration, intensity) International partnerships Australia – Australian version of Toolkit, 3 RCTs commissioned Chile – under development

Implications for research, policy and practice Research Further discussion about optimal analysis approaches Statistical Analysis Plans May explain some of the heterogeneity in meta-analyses (limits to precision) Importance of replication and meta-analysis More methodological exploration! Policy and Practice Need to communicate uncertainty estimates carefully as often dependent on analytic approach “What Works?” or “What’s Worked”? Communicate range as well as mean?

References Higgins, S. (2016) Meta-synthesis and comparative meta-analysis of education research findings: some risks and benefits Review of Education 4.1: 31–53. http://dx.doi.org/10.1002/rev3.3067 Higgins, S. & Katsipataki, M. (2016) Communicating comparative findings from meta-analysis in educational research: some examples and suggestions International Journal of Research & Method in Education 39.3 pp 237-254 http://dx.doi.org/10.1080/1743727X.2016.1166486 Kasim, A., Xiao, Z. & Higgins, S. (in preparation) eefAnalytics: A Package for Trial Data Analysis The R Journal Xiao Z., Kasim, A., Higgins, S.E. (2016) Same Difference? Understanding Variation in the Estimation of Effect Sizes from Educational Trials International Journal of Educational Research 77: 1-14 http://dx.doi.org/10.1016/j.ijer.2016.02.001

View of Durham City from the train station