National Evaluation of Specialty Selection Hywel Thomas and Celia Taylor On behalf of the NESS team: Ian Davison, Steven Field, Harry Gee, Janet Grant,

Slides:



Advertisements
Similar presentations
World Class Selection June 2008 North 51 - A team you can rely on The Psychology of Selection.
Advertisements

Martin Hart Assistant Director Education Case study on accreditation: the GMC’s perspective.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
Combining Test Data MANA 4328 Dr. Jeanne Michalski
Chapter 10 Decision Making © 2013 by Nelson Education.
Statistics of EBO 2010 Examination EBO General Assembly Sunday June 21st, 2010 (Tallin, Estonia) Danny G.P. Mathysen MSc. Biomedical Sciences EBOD Assessment.
1]Knoll, N., Burkert, S., & Schwartzer, R. (2006). Reciprocal support provision: Personality as a moderator? European Journal of Personality, 20,
Concept of Reliability and Validity. Learning Objectives  Discuss the fundamentals of measurement  Understand the relationship between Reliability and.
Effect of Selection Ratio on Predictor Utility Reject Accept Predictor Score Criterion Performance r =.40 Selection Cutoff Score sr =.50sr =.10sr =.95.
UKPHR Practitioner Registration East of England Public Health Practitioner Registration Scheme LAUNCH EVENT 5 th December 2014.
Correlational Designs
Research Methods in MIS
Chapter 7 Correlational Research Gay, Mills, and Airasian
Inferential statistics Hypothesis testing. Questions statistics can help us answer Is the mean score (or variance) for a given population different from.
Inferential Statistics
February  Study & Abstract StudyAbstract  Graphic presentation of data. Graphic presentation of data.  Statistical Analyses Statistical Analyses.
Kypros Kypri School of Medicine and Public Health University of Newcastle, Australia Injury Prevention Research Unit Department of Preventive & Social.
Work based assessment Challenges and opportunities.
Evaluating the Validity of NLSC Self-Assessment Scores Charles W. Stansfield Jing Gao Bill Rivers.
Designing an Evaluation of the Effectiveness of NIH’s Extramural Loan Repayment Programs.
Development version 19/06/ of 48 Effectiveness of a postural care training programme © 2012 Effectiveness of a postural care education programme.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Foundations of Recruitment and Selection I: Reliability and Validity
Self-reported cognitive and emotional effects and lifestyle changes shortly after preventive cardiovascular consultations in general practice Dea Kehler.
MGTO 324 Recruitment and Selections Validity II (Criterion Validity) Kin Fai Ellick Wong Ph.D. Department of Management of Organizations Hong Kong University.
Staffing & Selection © Nancy Brown Johnson 2000 Selection  Determining who will staff the organization.  Includes: interviewing, tests, weighing education.
L 1 Chapter 12 Correlational Designs EDUC 640 Dr. William M. Bauer.
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
CHAPTER 4 Employee Selection
Chapter Seven Measurement and Decision-Making Issues in Selection.
Tests and Measurements Intersession 2006.
Lab 5 instruction.  a collection of statistical methods to compare several groups according to their means on a quantitative response variable  Two-Way.
The Correlational Research Strategy
Power Point Slides by Ronald J. Shope in collaboration with John W. Creswell Chapter 12 Correlational Designs.
By Cao Hao Thi - Fredric W. Swierczek
th Annual Interdisciplinary Research Conference, Trinity College, Dublin Knowledge and practice of blood transfusion: a survey of nurses.
Part 5 Staffing Activities: Employment
Discriminant Analysis Discriminant analysis is a technique for analyzing data when the criterion or dependent variable is categorical and the predictor.
© Imperial College LondonPage 1 Applying for General Practice Training in England for August 2007 Professor Azeem Majeed Department of Primary Care & Social.
Hires 5 Offers 10 Interviews 40 Invites 60 Applicants 240 Adapted from R.H. Hawk, The Recruitment Function (New York: American Management Association,
BMH CLINICAL GUIDELINES IN EUROPE. OUTLINE Background to the project Objectives The AGREE Instrument: validation process and results Outcomes.
Self-assessment Accuracy: the influence of gender and year in medical school self assessment Elhadi H. Aburawi, Sami Shaban, Margaret El Zubeir, Khalifa.
Combining Test Data MANA 4328 Dr. Jeanne Michalski
Recruitment Statistics Overview of Recruitment 2007/08 Of the 38 vacancies advertised this year, 27 roles were filled and 7 roles were not.
Recruitment Statistics Total Number of Enquiries Made to Vacancies in 2006/07 In 2006/07 a total of 175 enquiries were made in to 21 academic.
Evaluating Risk Adjustment Models Andy Bindman MD Department of Medicine, Epidemiology and Biostatistics.
United Nations Workshop on Revision 3 of Principles and Recommendations for Population and Housing Censuses and Evaluation of Census Data, Amman 19 – 23.
© (2015, 2012, 2008) by Pearson Education, Inc. All Rights Reserved Chapter 11: Correlational Designs Educational Research: Planning, Conducting, and Evaluating.
Using propensity score matching to understand what works in reducing re-offending GSS Methodology Symposium Sarah French & Aidan Mews, Ministry of Justice.
Selecting Employees DeNotra Geddis April 11, 2005.
The Correlational Research Strategy Chapter 12. Correlational Research The goal of correlational research is to describe the relationship between variables.
Experimental Research Methods in Language Learning Chapter 12 Reliability and Reliability Analysis.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Developing a Hiring System Measuring Applicant Qualifications or Statistics Can Be Your Friend!
Topic #5: Selection Theory
School of Nursing Health Literacy Among Informal Caregivers of Persons With Memory Loss Judith A. Erlen, PhD, RN, FAAN; Jennifer H. Lingler, PhD, RN; Lisa.
1 Information Systems Use Among Ohio Registered Nurses: Testing Validity and Reliability of Nursing Informatics Measurements Amany A. Abdrbo, RN, MSN,
NHS Employers: Values Based Recruitment. ‘‘Values Based Recruitment is an approach which attracts and selects students, trainees or employees on the basis.
School of Allied Health Professions & School of Medicine, Health Policy & Practice LAMP A pragmatic unblinded randomised controlled trial and economic.
Correlates of the use of urgent care (UC) and quality of life in patients with 4 long-term physical conditions (CHD, asthma, diabetes and/or COPD). Authors:
Lecture 5 Validity and Reliability
Trena M. Ezzati-Rice, Frederick Rohde, Robert Baskin
MANA 4328 Dr. Jeanne Michalski
MANA 4328 Dennis C. Veit Measurement MANA 4328 Dennis C. Veit 1.
Calculating Sample Size: Cohen’s Tables and G. Power
MANA 4328 Dennis C. Veit Measurement MANA 4328 Dennis C. Veit 1.
The first test of validity
15.1 The Role of Statistics in the Research Process
UCLA Department of Medicine
Presentation transcript:

National Evaluation of Specialty Selection Hywel Thomas and Celia Taylor On behalf of the NESS team: Ian Davison, Steven Field, Harry Gee, Janet Grant, Andy Malins, Laura Pendleton and Elizabeth Spencer NESS was commissioned and funded by the Policy Research Programme in the Department of Health (Award number ). The views expressed are not necessarily those of the Department.

Background Specialty selection process is one of the hurdles on the way to consultant/GP principal posts 2009: 11,417 applicants for 6,580 entry-level posts (competition ratio 1.7 to 1) Selection became increasingly politically sensitive following MTAS Highlighted need for evolution and evaluation

Aims and scope of NESS To evaluate the first round of selection for specialty training in 2009 against four key criteria: – Acceptability – Fairness – Effectiveness: Validity and Reliability – Value for money 13 specialties included in the project Data collection primarily in 5 deaneries Did not obtain complete data for every specialty/deanery

Data Sources

Acceptability: “This selection process was fair” * * *

Fairness: Effect of personal characteristics on selection scores CharacteristicUnweighted mean coefficient across specialties N specialties in which coefficient statistically significant (p<0.05)/5 Male (vs. Female) Asian (vs. White) Other Ethnic group (vs. White) EU trained (vs. UK trained) Non-EU trained (vs. UK trained) FY2 post (vs. other post)0.101 Age in years Postgrad experience in years R squared0.29 (range 0.08 to 0.45)N/A Multiple linear regression analysis by specialty Standardised scores so comparability across specialties N=5 specialties and 1,553 candidates

Effectiveness: Predictive validity of shortlisting scores Selection ProcessUncorrectedCorrected Specialty 1 Deanery F Specialty 2 Deanery A Specialty 5 Deanery B Specialty 5 Deanery E Specialty 6 Deanery A Specialty 7 Deanery A0.48N/A Specialty 7 Deanery E0.56N/A Specialty 9 Deanery A0.17N/A Specialty 9 Deanery B Specialty 9 Deanery E Specialty 10 Deanery F Specialty 13 Deanery F Pearson correlation coefficients: uncorrected and corrected for restriction of range and unreliability of shortlisting scores where possible N=8 specialties, 13 selection processes and 2,411 candidates

Effectiveness: Reliability Internal Consistency: Cronbach’s alpha by station – N=10 specialties, 26 selection processes and 3,505 candidates – Range 0.35 to 0.83 – 10/26 (38%) in recommended range 0.7 to 0.9 Inter-rater reliability: Station-level absolute intra-class correlations – N=4 specialties, 4 selection processes and 395 candidates – Range 0.54 to 0.91 – 16/17 (94%) above recommended minimum of 0.7 Pass-Mark reliability (ignores sub-rules at station-level and only includes candidates attending interview) – N=5 specialties, 7 selection processes and 919 candidates – 12% to 55% of candidates within 1 SEM of appointment cut-off: raises concerns about fairness – 0% to 20% of candidates within 1 SEM of competency cut-off: raises concerns about competency

Pass-Mark reliability example

Value for Money Costing model developed ( Modified Brogden’s model to estimate cost-benefit: Cost-benefit depends on: – Selection process design – Predictive validity – Competition ratio – SD of training performance of candidates – Length of training – Drop-out rate – Number requiring extensions to training – Proportion unsuccessful candidates remaining in NHS Cost estimates for ST1 selection: £3.2m for hospital specialties (£800 per post) and £2.4m for GP (£900 per post) Cost-benefit estimates - compared to random selection - ranged from £78- 97m for hospital specialties and £15-20m for GP

Summary and implications for selection Largest study of specialty selection Did not obtain complete data – but no evidence of response bias High acceptability of selection processes by candidates and assessors Shortlist scores are a good predictor of selection scores Long-term follow-up is required on predictive validity, particularly to assess fairness (if scores are predictive then UK-trained candidates will make better trainees but need evidence) Inter-rater reliability was good – but potential collusion? Internal consistency and so pass-mark reliability could be improved: more stations with 1 assessor? Only one specialty had a formal standard setting process to identify competency cut-off Value for money could only be estimated – but suggests high returns to investment in selection Selection has continued to evolve since 2009 e.g. increase in nationally- coordinated selection processes