Times they are changing: Evaluating outreach in a new era

Slides:



Advertisements
Similar presentations
Introduction to Impact Assessment
Advertisements

Evaluating and Measuring Impact Presented by – Date –
Using Data to Measure and Report Program Impact Anne Basham, MFA MEMconsultants.
OECD/INFE High-level Principles for the evaluation of financial education programmes Adele Atkinson, PhD OECD With the support of the Russian/World Bank/OECD.
Good Evaluation Planning – and why this matters Presentation by Elliot Stern to Evaluation Network Meeting January 16 th 2015.
Monitoring and evaluation of carers’ services and projects Dr Andrea Wigfield - Associate Professor of Social Policy Centre for International Research.
ACADEMIC INFRASTRUCTURE Framework for Higher Education Qualifications Subject Benchmark Statements Programme Specifications Code of Practice (for the assurance.
Assessment in the early years © McLachlan, Edwards, Margrain & McLean 2013.
Monitoring Evaluation Impact Assessment Objectives Be able to n explain basic monitoring and evaluation theory in relation to accountability n Identify.
Implementation & Evaluation Regional Seminar ‘04 School Development Planning Initiative “An initiative for schools by schools”
Higher Education Access Tracker Helping member universities to target, monitor and evaluate outreach programmes Sharon Smith - Director of HEAT project.
Impact assessment framework
Evaluation methods and tools (Focus on delivery mechanism) Jela Tvrdonova, 2014.
Another New Framework Major Changes: No more satisfactory 2 strikes and you are out All criteria changed Very short notice No pre-inspection brief.
History PGCE Subject Development Panel Jan Feedback from Chief External Examiner 2013/14 KEY STRENGTHS: 1.Highly efficient and effective communication.
OECD/INFE Tools for evaluating financial education programmes Adele Atkinson, PhD Policy Analyst OECD With the support of the Russian/World Bank/OECD Trust.
QUICK OVERVIEW - REDUCING REOFFENDING EVALUATION PACK INTRODUCTION TO A 4 STEP EVALUATION APPROACH.
EVALUATION APPROACHES Heather Aquilina 24 March 2015.
1 Analysing the contributions of fellowships to industrial development November 2010 Johannes Dobinger, UNIDO Evaluation Group.
Evaluation To consider evaluation theory and apply it to the Welcome to Your Library project plans and begin to determine steps projects needs to take.
Primary.  There was a greater level of improvement in Literacy than Numeracy for both FSME and Non-FSME pupils.  Boys showed a greater level of.
The Coseley School A Co-operative Trust Closing the Gap Strategies – 2015/16 Believe, Achieve, Excel Closing the Gap Strategies – 2015/16 Believe, Achieve,
Developing a Project Proposal - SPROUT - ACTRAV-Turin.
Development Team Day 5a October Aim To explore approaches to evaluating the impact of the curriculum on pupil learning.
The Literacy and Numeracy Secretariat Le Secrétariat de la littératie et de la numératie October – octobre 2007 The School Effectiveness Framework A Collegial.
Advancing teaching: inspiring able learners every day Meeting the Challenge 14 th November 2012.
Supported by: Student Hubs is a registered charity in England and Wales, number Impact Measurement Proving and Improving your Social Impact.
Source: Patton, p1.. Program Evaluation: The systematic use of empirical information to assess and improve programs or policies.
Supporting Teacher Assessment - English. “This half-day meeting is part of a package available to purchase through our e-store. The package is designed.
What Difference Can Portfolio Make in Radiographer Work Practice
Evaluating the Quality and Impact of Community Benefit Programs
Self Evaluation in Partnership
Monitoring and Evaluation in Asset Based Approaches
Welcome - Pupil Premium
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 3. Session 9. Periodic data collection methods.
Gathering a credible evidence base
Staff meeting Monday 20th February 2017
Gender-Sensitive Monitoring and Evaluation
Gender-Sensitive Monitoring and Evaluation
LSRN Discussion Workshop: 24 November 2016
Key recommendations Successful components of physical activity interventions fall into three categories: Planning and developing physical activity initiatives.
Introduction to evaluating and measuring impact in career development Presented by – Date – Doubt, the essential preliminary of all improvement and.
Evaluation Emma King.
Managing for Results Capacity in Higher Education Institutions
April 21 Francesca Recanatini, WBI
New developments in the UK Higher Education
Still Building a Brighter Future for Young People through Sport
Conducting Efficacy Trials
Programme Review Dhaya Naidoo Director: Quality Promotion
Program Evaluation Essentials-- Part 2
Prepared by BSP/PMR Results-Based Programming, Management and Monitoring Presentation to Geneva Group - Paris Hans d’Orville Director, Bureau of Strategic.
Evaluation Jacqui McDowell.
Logic Models and Theory of Change Models: Defining and Telling Apart
Assessment Jargon The who, what, where, why, when, and how of writing end-of-the-year reports Sarah Todd Director of Institutional Research and Assessment.
General Notes Presentation length - 10 – 15 MINUTES
Introduction to Student Achievement Objectives
PROCESSES AND INDICATORS FOR MEASURING THE IMPACT OF EQUALITY BODIES
Building Knowledge about ESD Indicators
CATHCA National Conference 2018
PSHE at Key Stages 1 – 2 Guidance on Assessment, Recording and Reporting If it is not assessment for learning – why are we doing it? Assessment must.
WHAT is evaluation and WHY is it important?
Standard for Teachers’ Professional Development July 2016
Workshop 4 Planning next steps.
To RCT, or not RCT: that is the question
The importance of effective evaluation in widening participation
Changing the Game The Logic Model
University of Derby Progress to Success Evaluation Framework
Evaluating WP initiatives: Overcoming the Challenges
The National Professional Qualification (NPQ) An overview
WP Research and Evaluation Manager
Presentation transcript:

Times they are changing: Evaluating outreach in a new era Jo Astley University of Derby Emma Church EMWPREP emwprep.ac.uk

Session outline University of Derby Progress to Success framework Monitoring and evaluation Access and Participation Plans… what does it mean for monitoring and evaluation….and practitioners? Thoughts and discussion

Progress to Success (P2S) Framework

EMWPREP Evaluation Framework Targeting Monitoring Evaluation Tracking Targeting To Inform Monitoring To Evidence Evaluation To Improve Tracking To Influence

P2S Evaluation Framework- Targeting Targeting for P2S takes place in two mediums - school and individual level. UoD select schools from Derby City and Derbyshire where there is a high numbers of widening participation students and offer the opportunity to take part in an intensive and highly targeted programme to raise awareness of opportunities at university and raise attainment. UoD are able to assess eligibility of individuals with these schools based on geographical indicators by using our postcode checker. They are also able to use school intelligence in respect of students in receipt of pupil premium or eligible for FSM.

P2S Evaluation Framework- Monitoring

P2S Evaluation Framework- Monitoring: P2S 2017/18 Year in Brief

P2S Evaluation Framework- Monitoring: P2S 2017/18 Year in Brief

P2S Evaluation Framework: Evaluation: Complementary Methods Baseline Surveys To measure distance travelled and change over time Individual Event Evaluation Forms To record the success and impact of the events Teacher Questionnaires Quality control and view of the programme as a whole Interviews and Focus Groups Helps provide in-depth insights into participants understanding Reflective Diaries Allow participants to assess their own progress during and after taking part in interventions Attainment Data To ascertain evidence of effect on academic ability

P2S Evaluation Framework: Tracking Next year will be the first year that we will have had a full P2S cohort going through the EMWPREP monitoring process. It will be at that point that we can look at those students for who we have consent and will be able to track their responses to attitudinal and knowledge tests and assess the long term impact of those activities. Cohort Academic Year 2015/16 2016/17 2017/18 2018/19 2019/20 2020/21 2021/22 2022/23 2023/24 2024/25 2025/26 2026/27 2027/28 2028/29 1 Yr7   HE Ready HESA Data 2 3 4 5 6

Source: OfS

Filling in the gaps……. By clearly articulating what we expect to happen and hypothesising how we expect to generate this change, we can design evaluations that are much more clearly focused on expected outcomes (Crockford 68:2017)

Logic Models, theory of Change and Progression Framework: What are our objectives? Why are they our objectives? Evidence of student voice. What are we doing to achieve your objectives? What is the evidence that what we are doing works?

Theory of Change – Activity Level – filling in the gaps Problem Activity Aim Inputs Outputs - Activities Outputs -Participation Causal Mechanisms Outcome/Impact Short Term Medium Term Long Term There is a lack of awareness of HE amongst WP learners Y10 residential summer school To provide participants with an HE experience and all the information they require to make a well informed decision regarding participation University Staff WA, academic, services, ambassadors Budget -£10 K Planning Materials EMWPREP Evaluation Event over all Why Go Ice breaker sessions Taster sessions Graduation Disco Halls of Residence Motivational speaker Games night in halls Y10 WP students (meeting X criteria) It has been shown through XX sources that …… Raised awareness of HE Participation in HE Completion of HE programme

Practitioners….evaluation and monitoring Combining theory and practice Student voice – Participatory Action Research Realist evaluation methodologies Existing frameworks

Evidence, evidence and more evidence Description Evidence used Claims you can make Type 1: Narrative The evaluation provides a narrative and a coherent theory of change to motivate its selection of outreach activities in the context of a coherent outreach strategy Evidence of impact elsewhere and/or in the research literature on outreach effectiveness or from your existing evaluation results We have a coherent explanation of what we do and why Our claims are research-based Type 2: Empirical Research The evaluation collects data on outcomes and impact and reports evidence that those receiving an intervention have better results, though this does not establish any direct causal effect Quantitative and/or qualitative evidence of a pre/post treatment change or a treatment/non-treatment difference We can demonstrate that our interventions are associated with beneficial results Type 3: Causality The evaluation methodology provides evidence of a causal effect of an intervention Quantitative and/or qualitative evidence of a pre/post treatment change on a treated group relative to an appropriate control or comparison group using an appropriate and robust research design We believe our intervention causes improvement and can demonstrate the difference using a control or comparison group

Points for Discussion Challenges for practitioners? How can our tracking service help? How are institutions responding? Causality – the holy grail of access? Back to the drawing board? Collaboration and dissemination

A final word from the OfS……….. ‘There is a link between evidence and practice. It is generally not possible to have strong evidence where actual practice is weak. Stronger practice draws on and creates strong evidence. Therefore, to an extent, indicators of evidence are also indicators of practice’ (OfS Access and participation standards of evidence 2019)

References and useful links https://www.officeforstudents.org.uk/advice-and-guidance/promoting-equal-opportunities/evaluation-and-effective- practice/standards-of-evidence-and-evaluation-self-assessment-tool/evaluation-self-assessment-tool/ - Self Assessment tool https://www.officeforstudents.org.uk/media/6971cf8f-985b-4c67-8ee2-4c99e53c4ea2/access-and-participation-standards-of- evidence.pdf - Access and participation standards of evidence https://www.officeforstudents.org.uk/media/f2424bc6-38d5-446c-881e-f4f54b73c2bc/using-standards-of-evidence-to-evaluate- impact-of-outreach.pdf - Using standards of evidence to evaluate impact of outreach https://www.officeforstudents.org.uk/publications/regulatory-notice-1-access-and-participation-plan-guidance/ OfS Regulatory notice 1 APP guidance Crockford J 2017 Five years of WPREU: Critical reflections of practice in widening participation and student success accessed June 2019 https://www.sheffield.ac.uk/apse/wp/wpreubook Miller M 2017 Five years of WPREU: Critical reflections of practice in widening participation and student success accessed June 2019 https://www.sheffield.ac.uk/apse/wp/wpreubook

Any Questions? Jo Astley Emma Church Widening Access Manager J.Astley@derby.ac.uk Emma Church EMWPREP Coordinator E.Church@lboro.ac.uk