Analysing educational trials: the Education Endowment Foundation Archive Steve Higgins, Adetayo Kasim, ZhiMin Xiao, with Nasima Akhter, Ewoud De Troyer,

Slides:



Advertisements
Similar presentations
A Spreadsheet for Analysis of Straightforward Controlled Trials
Advertisements

Povertyactionlab.org Planning Sample Size for Randomized Evaluations Esther Duflo J-PAL.
© Institute for Fiscal Studies Evaluation design for Achieve Together Ellen Greaves and Luke Sibieta.
Exploring Research-Led Approaches to Increasing Pupil Learning Steve Higgins School of Education, Durham University Addressing.
Building Evidence in Education: Workshop for EEF evaluators 2 nd June: York 6 th June: London
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
What role should probabilistic sensitivity analysis play in SMC decision making? Andrew Briggs, DPhil University of Oxford.
Using evidence to raise the attainment of children facing disadvantage James Richardson Senior Analyst, Education Endowment Foundation 1 st April 2014.
Session 4: Analysis and reporting Managing missing data Rob Coe (CEM, Durham) Developing a statistical analysis plan Hannah Buckley (York Trials Unit)
15 de Abril de A Meta-Analysis is a review in which bias has been reduced by the systematic identification, appraisal, synthesis and statistical.
Joining the dots Closing the Gap Getting a better understanding of the data 1 Mike Treadaway, FFT.
A randomised controlled trial to improve writing quality during the transition between primary and secondary school Natasha Mitchell, Research Fellow Hannah.
Statistics By Z S Chaudry. Why do I need to know about statistics ? Tested in AKT To understand Journal articles and research papers.
Today Concepts underlying inferential statistics
2 Accuracy and Precision Accuracy How close a measurement is to the actual or “true value” high accuracy true value low accuracy true value 3.
Making all research results publically available: the cry of systematic reviewers.
School of Education Archive Analysis: On EEF Trials Adetayo Kasim, ZhiMin Xiao, and Steve Higgins.
Using research to get the best value from the Pupil Premium Steve Higgins, School of Education, Durham National.
Research evidence and effective use of the Pupil Premium Professor Steve Higgins, School of Education, Durham
Addressing educational disadvantage, sharing evidence, finding out what works Camilla Nevill Evaluation Manager.
Planning high quality, evidence based provision to meet the needs and achieve the outcomes How do you know what works?
Building Evidence in Education: Conference for EEF Evaluators 11 th July: Theory 12 th July: Practice
The Pupil Premium: Using Evidence to Narrow the Gap Robbie Coleman 7 th July 2014
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 6.
META-ANALYSIS, RESEARCH SYNTHESES AND SYSTEMATIC REVIEWS © LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON.
Impact of two teacher training programmes on pupils’ development of literacy and numeracy ability: a randomised trial Jack Worth National Foundation for.
Developing teaching as an evidence informed profession UCET Annual Conference Kevan Collins - Chief Executive
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
The rise of RCTs in education: statistical and practical challenges Kevan Collins.
Approaches to quantitative data analysis Lara Traeger, PhD Methods in Supportive Oncology Research.
Talk Boost A targeted intervention for 4-7 year olds with language delay Wendy Lee Professional Director, The Communication Trust Mary Hartshorne Head.
Statistical Analysis Plans EEF Evaluators’ Conference 2016 Dr Ben Styles Head of NFER’s Education Trials Unit.
Methods of Presenting and Interpreting Information Class 9.
Meta-analysis Overview
Evaluation in Education: 'new' approaches, different perspectives, design challenges Camilla Nevill Head of Evaluation, Education Endowment Foundation.
2016 Primary Assessment Update 27th September 2016
I-squared Conceptually, I-squared is the proportion of total variation due to ‘true’ differences between studies. Proportion of total variance due to.
Analysing the Primary RAISE
How to give research results back to teachers?
Effect Sizes.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Analysis for Designs with Assignment of Both Clusters and Individuals
Professor Steve Higgins, School of Education, Durham University
Exercise 2-Effect Size Coding
Lurking inferential monsters
Challenges arising from the analysis of randomized trials in education
The English RCT of ‘Families and Schools Together’
UOG Journal Club: October 2016
Statistical Core Didactic
How to read a paper D. Singh-Ranger.
What is the EEF Data Archive?
Group Comparisons What is the probability that group mean differences occurred by chance? With Excel (or any statistics program) we computed p values to.
EEF Archive analysis overview
Heterogeneity and sources of bias
Evidence in Action: Using Research to Narrow the Gap Eleanor Stringer
Statistical Analysis Plan review
Analyzing Intervention Studies
Power, Sample Size, & Effect Size:
EVAAS Overview.
Systematic review and meta-analysis
Gerald Dyer, Jr., MPH October 20, 2016
Brahm Fleisch Research supported by the Zenex Foundation October 2017
Workshop 4 Planning next steps.
SAMPLING AND STATISTICAL POWER
The EEF Data Archive: Plans & Priorities
Applying to the EEF for funding: What are we looking for
Publication Bias in Systematic Reviews
CCSSO National Conference on Student Assessment June 21, 2010
Sample Sizes for IE Power Calculations.
Meta-analysis, systematic reviews and research syntheses
Presentation transcript:

Analysing educational trials: the Education Endowment Foundation Archive Steve Higgins, Adetayo Kasim, ZhiMin Xiao, with Nasima Akhter, Ewoud De Troyer, Quentin Dhumeaux, Mengchu Li s.e.higgins@durham.ac.uk @profstig 'Bristol Conversations in Education' seminar series, School of Education, Bristol University 24th October, 2018

The EEF Archive aims: 1. To track the impact of EEF projects longitudinally (by linking to the National Pupil Database) 2. To compare and group the effects of different projects so to draw lessons about which approaches to teaching and learning are most effective generally, and which are effective for different groups (such as FSM) 3. To answer methodological questions about research design, analysis and outcome measurement in educational trials 4. To investigate the cumulative impact of projects on schools and pupils; and 5. To understand better EEF’s target group of disadvantaged pupils.

The Archiving Process Independent evaluator submits data from EEF project to FFT FFT re-match with National Pupil Database (NPD) Batch transfers to Durham every six months New NPD data added when available (annually) 65 projects to date (out of 135 commissioned) About 750,000 pupils EEF Archive small grants – DfE/ONS

Principles and caveats Principles Independent evaluator’s peer-reviewed published estimate is always the official EEF finding for impact on attainment Exploratory analysis to inform EEF’s evaluation strategy Help to explain variation in trial impact Agreed process should errors be identified Caveats Can’t always match N (4/39 projects greater than 10% difference)* NPD data removed and re-matched Additional covariates may alter the N and alter the effect size Some raw scores transformed Gain score distribution may differ from post-test * ‘included in final analysis’ tag now required on deposit with FFT

Durham process and progress Initial exploration Development of analytic models Different effect size calculation (Cohen’s d, Hedges’ g, Glass’ Δ), Post-test only comparisons (δ) Gain score Ordinary least squares (OLS) Multi-level models Explore variances within, between, total Frequentist Bayesian Design specific: individual randomisation, cluster, multi-site Initial exploration Development of analytic models Different effect size calculation (Cohen’s d, Hedges’ g, Glass’ Δ), Post-test only comparisons (δ) Gain score Ordinary least squares (OLS) Multi-level models Explore variances within, between, total Frequentist Bayesian Design specific: individual randomisation, cluster, multi-site R package ‘EEFanalytics’ https://CRAN.R-project.org/package=eefAnalytics Control for: Baseline School/ clustering Design Xiao Z., Kasim, A., Higgins, S.E. (2016) Same Difference? Understanding Variation in the Estimation of Effect Sizes from Educational Trials International Journal of Educational Research 77: 1-14 http://dx.doi.org/10.1016/j.ijer.2016.02.001

Initial findings 17 EEF projects Four analytic models Findings Results converge in larger, unproblematic trials Point estimates and estimates of precision vary when trials are problematic (e.g. testing issues, randomisation), when design and analysis not matched; with different co-variates added; when outcomes are different or transformed (e.g. z-scores) Results tend to diverge if ICC ≥ 0.2 and there are few clusters/schools MLM total variance ‘most conservative’ model (wider Confidence Intervals) Bayesian estimates identical but more precise (narrower Credible Intervals)

Variation 56 outcomes from 39 projects 32 outcomes small difference ≤ 0.05 19 between 0.06 and 0.10 2 between 0.11 and 0.19 2 large difference > 0.20 (Discover Summer School, Mind the Gap) Mean difference is zero

Variation in estimates Stem Leaf -0.2 -0.1 -0.0 0.0 5 1 1 1 1 2 3 4 4 4 4 4 4 4 5 5 5 5 6 6 6 7 7 7 7 7 8 9 9 0 0 0 0 1 1 1 1 2 3 4 4 4 5 5 5 6 7 7 8 8 8 0.1 1 5 7 7 8 0.2 4 Stem Leaf -0.2 -0.1 -0.0 0.0 5 1 1 1 1 2 3 4 4 4 4 4 4 4 5 5 5 5 6 6 6 7 7 7 7 7 8 9 9 0 0 0 0 1 1 1 1 2 3 4 4 4 5 5 5 6 7 7 8 8 8 0.1 1 5 7 7 8 0.2 4 Stem and leaf plot of differences between Evaluator and archive MLM total variance estimates for 56 primary outcomes across 39 projects

Overview of variation

Sample of evaluator models Xiao Z., Kasim, A., Higgins, S.E. (2016) Same Difference? Understanding Variation in the Estimation of Effect Sizes from Educational Trials International Journal of Educational Research 77: 1-14 http://dx.doi.org/10.1016/j.ijer.2016.02.001

Convergence: Text Now Pre-test imbalance

Clustering: Maths Mastery Secondary Point estimates similar, CIs vary

Divergence: Fresh Start Large pre-test imbalance Post-ANCOVA models correct for this Evaluator gain estimate much higher Post/gain differences

Other explorations What extent of imbalance at baseline can be adequately addressed with a post-test analysis (Lord’s Paradox)? Treating multisite trials as cluster randomised trials and using Hedges’ (2007) formula can be overly conservative Exploring individualized treatment effects/ updating binomial effect size display (Rosenthal & Rubin 1982) How best to deal with missing data - different imputation methods produce different effect size estimates and can also introduce bias Subgroup or interaction analysis – when is each better?

Thorny issues and reflections Thorny issues p-values and uncertainty estimates Messages – project level & archive level consistency? Means or distributions? (cf. Deaton & Cartwright (2016) on RCTs & Average Treatment Effects) Replication and direct comparisons? Reflections What is acceptable analytic variation? What is the best estimate of the ‘average treatment effect’? How much should evaluators be constrained for comparability? With intention to treat approach, sample sizes predicated on minimum necessary to detect a probably overestimated EEF Toolkit effect size, and using MLM total variance are we being too conservative?

Publications Kasim, A., Xiao Z., Higgins, S.E & De Troyer, E. (2014) eefAnalytics [Computer software] https://CRAN.R-project.org/package=eefAnalytics Higgins, S. (2016) Meta-synthesis and comparative meta-analysis of education research findings: some risks and benefits Review of Education 4.1: 31–53. http://dx.doi.org/10.1002/rev3.3067 Higgins, S. & Katsipataki, M. (2016) Communicating comparative findings from meta-analysis in educational research: some examples and suggestions International Journal of Research & Method in Education 39.3 pp 237-254 http://dx.doi.org/10.1080/1743727X.2016.1166486 Xiao Z., Kasim, A., Higgins, S.E. (2016) Same Difference? Understanding Variation in the Estimation of Effect Sizes from Educational Trials International Journal of Educational Research 77: 1-14 http://dx.doi.org/10.1016/j.ijer.2016.02.001 Xiao, Z. Higgins, S. & Kasim, A (2017) An Empirical Unravelling of Lord's Paradox The Journal of Experimental Education http://dx.doi.org/10.1080/00220973.2017.1380591 Xiao, Z., & Higgins, S. (2018). The power of noise and the art of prediction. International Journal of Educational Research, 87, 36-46. http://dx.doi.org/10.1016/j.ijer.2017.10.006 Higgins, S. (2018) Improving Learning: Meta-analysis of Intervention Research in Education Cambridge: Cambridge University Press http://dx.doi.10.1017.9781139519618