Building an Employment Data Lab

Slides:



Advertisements
Similar presentations
Paul Smith Office for National Statistics
Advertisements

Challenges in evaluating social interventions: would an RCT design have been the answer to all our problems? Lyndal Bond, Kathryn Skivington, Gerry McCartney,
OECD/INFE High-level Principles for the evaluation of financial education programmes Adele Atkinson, PhD OECD With the support of the Russian/World Bank/OECD.
Estimating net impacts of the European Social Fund in England Paul Ainsworth Department for Work and Pensions July 2011
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Good Evaluation Planning – and why this matters Presentation by Elliot Stern to Evaluation Network Meeting January 16 th 2015.
Justice Data Lab: Facing the Third Sector How can we develop the capacity of third sector organisations to engage with data? Scottish Universities Insight.
PAI786: Urban Policy Class 2: Evaluating Social Programs.
Designing an Evaluation of the Effectiveness of NIH’s Extramural Loan Repayment Programs.
Criteria for Assessing The Feasibility of RCTs. RCTs in Social Science: York September 2006 Today’s Headlines: “Drugs education is not working” “ having.
OECD/INFE Tools for evaluating financial education programmes Adele Atkinson, PhD Policy Analyst OECD With the support of the Russian/World Bank/OECD Trust.
Welfare Reform and Lone Parents Employment in the UK Paul Gregg and Susan Harkness.
Centre for Market and Public Organisation Using difference-in-difference methods to evaluate the effect of policy reform on fertility: The Working Families.
Evidencing Outcomes Ruth Mann / George Box Commissioning Strategies Group, NOMS February 2014 UNCLASSIFIED.
1 1 How to reduce the reporting burden whilst still obtaining high quality data Two practical examples from Norwegian financial markets statistics Ole.
Using propensity score matching to understand what works in reducing re-offending GSS Methodology Symposium Sarah French & Aidan Mews, Ministry of Justice.
The challenge of evidence-based policy Mediation and Alternative Dispute Resolution – Outcomes and Impacts Swansea University 14 th February 2013 Professor.
1 Joint meeting of ESF Evaluation Partnership and DG REGIO Evaluation Network in Gdańsk (Poland) on 8 July 2011 The Use of Counterfactual Impact Evaluation.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Chapter 10 Confidence Intervals for Proportions © 2010 Pearson Education 1.
People lives communities Supported employment for disabled people Commissioning and Contracting Training Conference 12 September 2014 Rich Watts, NDTi.
Looking for statistical twins
High Level Policy on Means Testing
Measuring Impact – what can we learn from recent DWP evaluations?
Employer retirement pension schemes
HEALTH ECONOMICS BASICS
The inspection of local areas effectiveness in identifying and meeting the needs of children and young people who have special educational needs and/or.
Patricia M. Alt, Ph.D. Dept. of Health Science Towson University
An agency of the Office of the Secretary of Education and the Arts
Using insight to support tribunals
Dr Kieran Fenby-Hulse & Dr Rebekah Smith McGloin
The Costs of Homelessness: Politics, Advocacy and Research
1 The roles of actuaries & general operating environment
AF1: Thinking Scientifically
How to read a paper D. Singh-Ranger.
The NICE Citizens Council and the role of social value judgements
Leaving no one behind: The value for money of disability-inclusive development Glasgow, 1st February 2017.
Skills Escalator Pilot and ESF proposal Thursday 23 April 2015.
Matching Methods & Propensity Scores
Variant Projections.
Statistics and Data Analysis
Matching Methods & Propensity Scores
Sabrina Iavarone Senior User Services Officer
Sue Todd Department of Mathematics and Statistics
ESF EVALUATION PARTNERSHIP MEETING Bernhard Boockmann / Helmut Apel
Development Impact Evaluation in Finance and Private Sector
Research Update GERI May 2010.
The Use of Counterfactual Impact Evaluation Methods in Cohesion Policy
Matching Methods & Propensity Scores
Implementation Challenges
Skills Escalator Pilot and ESF proposal Thursday 23 April 2015.
REAL (working with Singizi) Presentation to NSA 3rd/4th August
III. Practical Considerations in preparing a CIE
Program Evaluation, Archival Research, and Meta-Analytic Designs
Statistics beyond the National Level –Regional Experiences
Overview of the Children and Families Act 2014
Evaluating Impacts: An Overview of Quantitative Methods
Meeting the emerging needs for statistics on migrants:
Class 2: Evaluating Social Programs
A Multi-disciplinary Perspective on Decision-making and Creativity:
Sampling for Impact Evaluation -theory and application-
Class 2: Evaluating Social Programs
Counterfactual Impact Analysis applied in the ESF-Evaluation in Austria (period ) Contribution to the Expert-Hearing: Member States Experiences.
Andrew Jenkins and Rosalind Levačić
Monitoring and Evaluating FGM/C abandonment programs
Task Force on Small and Medium Sized Enterprise Data (SMED)
ISABEL NAYLON ESF EVALUATION PARTNERSHIP MEETING 13 NOVEMBER 2013
Discussion of Baugh (2015) “What happens when payday borrowers are cut off from payday lending? A natural experiment” Brian T. Melzer Kellogg School of.
The Application of Statistical Matching to the 2010 ESF Leavers Survey
Estimating net impacts of the European Social Fund in England
Presentation transcript:

Building an Employment Data Lab Mike Daly, Central Analysis & Science Strategy UK Department of Work and Pensions September 2019

Outline of talk Background – why do we need a Data Lab? The nature of the problem – why providers cannot do it for themselves, nor DWP for them with existing approach The proposed solution – an ‘Employment Data Lab’ The advantages of the approach The challenges we face Possible future developments

1) Background – why do we need a Data Lab? Employment support/labour market programmes increasingly provided outside DWP. Large number of providers - Employment Related Services Association has 250 members – 77% not for profit A variety of funding sources including: National Lottery Community Fund Other charities Local government Huge appetite to know ‘what works’: Through a natural desire to provide the best support possible But also for funders, who often demand a ‘robust evaluation’ And for DWP, to build the evidence base

2) The problem (a) why providers cannot do it for themselves Providers may be willing - even eager - to provide evaluation evidence but: Collecting data on outcomes is expensive, especially when extended beyond the end of a programme Even if providers collected their own data, it would almost certainly be inconsistent They have no access to non-participants to create a counterfactual Even if they did, they do not have the expertise to carry out an impact assessment – nor the budget to pay for it But surely, DWP have access to administrative data, and extensive experience in using this for impact assessments, so..?

2) The problem (b) why usual DWP approach cannot be used When we evaluate our own programmes we rely on: No legal issues – because we fund the programme, it is ‘our’ data We know who participants are, because of the funding mechanism Which includes a direct link into our systems so adding data such as labour market history is simple We have data on non-participants A proportion of the resource allocated to the programme helps support the evaluation work But for an independently funded programme these do not apply – in particular: A legal basis for sharing the data needs to be found We simply do not have the resource to tackle more than a handful

3) The proposed solution A DWP service – the Employment Data Lab - to enable the evaluation of employment related interventions carried out by third party organisations We use administrative data only – DWP records on benefit receipt, HMRC records on employment – no new data collection Providers send list of participants, and participation dates, to DWP We identify them in our records using fuzzy matching, and attach work and benefits data From this we can report on outcomes in a completely consistent way, and also compare participants to non-participants (e.g. older than average, with less of a work history) Where appropriate, we select a comparison group using propensity score matching and give net impact estimates – again, in a completely consistent way Providers get a report with aggregate results – which is published on our web site

4) The key advantages of the approach For the providers, it is a free service It is highly secure - no individual level data leaves DWP, nor is any provider data retained by DWP Results are robust – using best practice in application of psm, with appropriate caveats Outputs are consistent – allowing meta-analysis as more reports accumulate, adding further value Where impact estimates cannot be produced or are inconclusive, the outcomes data alone has considerable value We are building on a model operated successfully for 6 years by Ministry of Justice, with over 200 separate reports [1] We are also building on substantial DWP experience in the use of psm [2]

5) The key challenges (1) Testing the Conditional Independence Assumption. How plausible is it that the administrative data can control for all relevant factors? Somewhat conflicting evidence on this: Caliendo et al [3] – using additional data collected through a survey does not materially affect the results – even though it is associated with selection Dolton et al [4] on the other hand found that it made a noticeable difference More obvious where selection is explicitly on unobserved variables – e.g. interventions aimed specifically at drug users We expect to rely to a fair extent on human consideration of process by which participants are selected, and the plausibility of the assumption that available data controls for that. Can we do any better? How useful are Rosenbaum bounds? Ministry of Justice routinely use them; we have, to date, not done so Are there any other approaches to assessing the CIA? Using fuzzy matching: Optimising matching algorithms to balance false positives and false negatives Understanding the analytical implications of bias in the matching process

5) The key challenges (2) – detailed implementation Resources are limited – we need to make the process as automatic as possible. Which decisions can be automated, which need human (and expert) intervention? How can we ensure that the default decisions we make, and our processes for making decisions in individual cases, represent accepted best practice, as far as possible Communicating the limitations of the results. Broadly speaking we want people to understand the difference between: (a) the results are probably broadly OK, even though we can’t be certain; (b) we’ve got real doubts – use with great caution; (c) we are not including the impact results because they are too unreliable We need to strike a balance between something which is accessible to, and usable by non-specialist readers, while still standing up to academic scrutiny Setting the right lower bound for sample size – we do not want to waste resource, or damage our credibility, on reports where we cannot provide an impact estimate, or only one with very wide confidence intervals. But equally, we do not want to turn away all but the largest providers Ensuring that we retain the requisite aggregates and statistics to support meta-analysis, given that we cannot retain the individual data

6) Future (possible) developments Using additional data – e.g. education, criminal justice, health – to improve matching, and to present a wider range of outputs Using alternative methods – instrumental variables, regression discontinuity for example. Key constraint here is that these are unlikely to be applicable in enough cases to be practical Reviewing the possibility of retaining data in some way, so as to allow longer-term follow-up, or analysis of multiple participations Considering interactions with other provision – e.g. does prior participation in a national DWP-funded programme lead to increased or reduced impact

References [1] see https://www.gov.uk/government/publications/justice-data-lab [2] See for example: Ainsworth and Marlow, 2011; Early impacts of the European Social Fund 2007-2013; DWP In House Report No 3 DWP, 2012; Impacts and costs and benefits of the Future Jobs Fund;   https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/223120/impacts_costs_benefits_fjf.pdf Anderson et al, 2004; Work-based Learning for Adults: an evaluation of labour market effects; DWP Working Age report 187 [3] Caliendo, Mahlstedt & Mitnik; 2017; Unobservable, but Unimportant? The Influence of Personality Traits (and Other Usually Unobserved Variables) for the Estimation of Treatment Effects; Labour Economics 46 (2017); 14-25 [4] Dolton, Smith & Azevedo; 2011; The Impact of the UK New Deal for Lone Parents on Benefit Receipt ; IZA Discussion Paper 5491