Standard Setting for a Performance-Based Examination for Medical Licensure Sydney M. Smee Medical Council of Canada Presented at the 2005 CLEAR Annual.

Slides:



Advertisements
Similar presentations
Knowledge Dietary Managers Association 1 PART II - DMA Certification Exam Blueprint and Exam Development-
Advertisements

Assessment of Initial Clinical Competence for NP Regulation: The Development of an Objective Structured Clinical Examination Marie Napolitano RN, PhD,
Principles of Standard Setting
CLEAR 2008 Annual Conference Anchorage, Alaska Strategies and Outcomes of Pre-testing for Performance-Based Assessments Carol OByrne, PEBC.
Assessment of the Cooperative Education Program in Relationship to the Program Outcomes of the Undergraduate Nursing and UHM Cooperative Education Programs.
OSCEs Kieran Walsh OSCE Means “objective structured clinical examination” Assesses competence Developed in light of traditional assessment methods.
“Scoring an Oral Simulation Exam” Elizabeth A. Witt, Ph.D. American Board of Emergency Medicine Presented at the 2005 CLEAR Annual Conference September.
Copyright © 2012 Pearson Education, Inc. or its affiliate(s). All rights reserved
PRIOR LEARNING ASSESSMENT RESEARCH PROJECT IN NURSING Mount Royal College, Calgary, Alberta Purpose of the Research: To assess the impact of.
M AKING A PPROPRIATE P ASS- F AIL D ECISIONS D WIGHT H ARLEY, Ph.D. DIVISION OF STUDIES IN MEDICAL EDUCATION UNIVERSITY OF ALBERTA.
S tructured O bjective C linical E xamination SOC E.
Advanced Topics in Standard Setting. Methodology Implementation Validity of standard setting.
Division of medical and dental education -update on OSCEs Rhoda MacKenzie.
Medical school attendedPassing grade Dr JohnNorthsouth COM (NSCOM)80% Dr SmithEastwest COM (EWCOM)50% Which of these doctors would you like to treat you?
Workplace-based Assessment. Overview Types of assessment Assessment for learning Assessment of learning Purpose of WBA Benefits of WBA Miller’s Pyramid.
Setting Performance Standards Grades 5-7 NJ ASK NJDOE Riverside Publishing May 17, 2006.
Graduate Program Assessment Report. University of Central Florida Mission Communication M.A. Program is dedicated to serving its students, faculty, the.
An overview of Assessment. Aim of the presentation Define and conceptualise assessment Consider the purposes of assessment Describe the key elements of.
An overview of Assessment. Aim of the presentation Define and conceptualise assessment Consider the purposes of assessment Describe the key elements of.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Standard Setting Inclusive Assessment Seminar Marianne.
Standard Setting Different names for the same thing Standard Passing Score Cut Score Cutoff Score Mastery Level Bench Mark.
AGS Annual Meeting May 2005 Converting the Geriatric Functional Assessment Standardized Patient Instructor …into an OSCE Karen E. Hall, M.D., Ph.D. Clinical.
Setting Alternate Achievement Standards Prepared by Sue Rigney U.S. Department of Education NCEO Teleconference March 21, 2005.
Assessment of Clinical Competence in Health Professionals Education
Standard setting Determining the pass mark - OSCEs.
Comparison: Traditional vs. Outcome Project Evaluative Processes Craig McClure, MD Educational Outcomes Service Group University of Arizona December 2004.
Training the OSCE Examiners
Triple C Competency-based Curriculum: Implications for Family Medicine Residency Programs.
Assessment Tools. Contents Overview Objectives What makes for good assessment? Assessment methods/Tools Conclusions.
Core Competency Assessment in Emergency Medicine from Design to Implementation National Hispanic Medical Conference Christian Arbelaez, MD, MPH Associate.
Standard Setting Methods with High Stakes Assessments Barbara S. Plake Buros Center for Testing University of Nebraska.
What is Happening to the National Board Dental Examination? Kari Cunningham ASDA Representative to the JCNDE Western Regional ASDA Meeting October 30,
Metropolis Pre-Conference Workshop Regulatory Bodies: Integration Partners and Gatekeepers YVES LAFORTUNE Director, Evaluation Bureau 12 March 2014 Hilton.
Quality in language assessment – guidelines and standards Waldek Martyniuk ECML Graz, Austria.
1 Wallace Judd APA Testing Standards: RFC
1 Establishing A Passing Standard Paul D. Naylor, Ph.D. Psychometric Consultant.
Classroom Assessments Checklists, Rating Scales, and Rubrics
R 3 P Colloquium American Board of Pediatrics Jan. 31 – Feb. 2, 2007 The Past, Present and Future Assessments of Clinical Competence A Canadian Perspective.
Sterling Practices in Design & Scoring of Performance-Based Exams #156 F. Jay Breyer Presented at the 2005 CLEAR Annual Conference.
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
Student assessment AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
1 Producing Your Assessment Question Mark “Software for creating and delivering assessments with powerful reports”  Copyright 2000 QuestionMark. All.
Patricia A. Mahoney, MSN, RN, CNE
Issues in Selecting Assessments for Measuring Outcomes for Young Children Issues in Selecting Assessments for Measuring Outcomes for Young Children Dale.
Dr. Paul Townsend, D.C. Director of Practical Testing, Research & Development NBCE Part IV Practical Examination Duties of a Part IV Examiner Dr. LeRoy.
Developing an Assessment System B. Joyce, PhD 2006.
Assessment Tools.
School of Clinical Medicine School of Clinical Medicine UNIVERSITY OF CAMBRIDGE UK Council problems with OSCE assessment Jonathan Silverman 2012.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Catholic College at Mandeville Assessment and Evaluation in Inclusive Settings Sessions 3 & /14/2015 Launcelot I. Brown Lisa Philip.
Chapter 6: High-Leverage Practice 1: Formative Evaluation.
EDUCATION NIGHT USMLE A three-step examination for medical licensure in the United States Step 1 (Basic Sciences)  System and Process  322 multiple.
Education Night USMLE A three-step examination for medical licensure in the United States Step 1 (Basic Sciences)  System and Process  308 multiple.
MRCGP The Clinical Skills Assessment January 2013.
Proposed End-of-Course (EOC) Cut Scores for the Spring 2015 Test Administration Presentation to the Nevada State Board of Education March 17, 2016.
INTRODUCTION TO ASSESSMENT METHODS USED IN MEDICAL EDUCATION AND THEIR RATIONALE.
Presented by Dr Safeera Hussainy OSCEology A primer on performance-based teaching, learning & assessment in pharmacy.
S tructured O bjective C linical E xamination P ractical.
Empathy in Medical Care Jessica Ogle (D
Clinical Assessment Dr. H
Challenges and options in assessment
Criterion Referencing Judges Who are the best predictors?
Father Muller Medical College & Hospital, Mangalore, Karnataka.
Medical Council of Canada Exam Changes
Assessment 101 Zubair Amin MD MHPE.
Competency and Performance: What Should We Measure and How
Assessment of Clinical Competencies
Approach to OSCES.
Evaluation Rating Forms
Purpose of this presentation
Presentation transcript:

Standard Setting for a Performance-Based Examination for Medical Licensure Sydney M. Smee Medical Council of Canada Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona MCC Qualifying Examination Part II OSCE format - 12 short stations ●5 or 10 minutes per patient encounter ●Physicians observe and score performance Required for medical licensure in Canada Prerequisites ●Passed MCCQE Part I (Knowledge & Clinical reasoning) ●Completed 12 months of post-graduate clinical training Pass/Fail criterion-referenced examination Multi-site administration - twice per year Overall fail rate 10%-30% Implemented 1992

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Why do it? Requested by licensing authorities, largely in response to two issues: ●Increase in complaints, many centered around communication skills. ●Public accountability - OSCE to serve as an “audit” of training of all candidates seeking licensure in Canada.

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Blueprint Considerations Four domains *History-taking *Patient Interaction *Physical Examination *Management Multi-disciplinary / multi-system content Patient demographics Two formats * 5+5 couplets & 10 minute Each case based on an MCC Objective

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Standard for MCCQE Part II Acceptably competent for entry to independent practice Conjunctive standard ●Pass by total score AND ●Pass by minimum number of stations High performance in a few stations does not compensate for overall poor performance Just passing enough stations does not compensate for overall poor performance

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Translating a Standard to a Pass Mark Pilot exam: Ebel method ●Items rated for relevance and importance ●Pass based on most relevant and important items ●Failed 40% First two administrations: Angoff method ●Estimated score for the minimally competent candidate ●Pass based on average of estimates per instrument ●Pass marks varied more than the test committee liked ●Test committee did not like the task 1994: Adopted borderline group method

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Physicians as Scorers Three Assumptions: Clinicians do not require training to judge candidate behaviour according to checklists for basic clinical skills Most clinicians can make expert judgments about candidate performance Being judged by clinicians is vital for a high-stakes examination

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Physicians as Standard Setters

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Global Rating Question Did the candidate respond satisfactorily to the needs/problems presented by this patient? Borderline Satisfactory Good Excellent Borderline Unsatisfactory Unsatisfactory Inferior

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Numbers....  1,000-2,200 candidates per administration Examiners each observe candidates examiners per case Number of candidates identified as borderline per case ranges from  Collect >99% of data for global rating item

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Modified Borderline Group Method Examiners (content experts) identify borderline candidates based on the 6-point scale Scores of borderline candidates define performance that “describes” the pass standard Examiner judgments are translated into a pass mark by taking the mean score for the borderline candidates for each case

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Pass Marks by Case Across Exams Challenge to assess pass marks over multiple administrations ●Scoring instruments are revised post-exam ●Rating scale items have been revised ●Rating scale items have been added to cases As competency and difficulty of cases changes, so do cut scores

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Setting Total Exam Pass Mark Pass marks for cases are summed Add one standard error of measure (  3.2% ) Pass mark falls between 1 to 1.5 SD below mean score Station performance is reviewed by Central Examination Committee ●Then the standard for the number of stations passed is set Standard has been 8/12 since 2000

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Outcomes 15,331 candidates became eligible in 2000 – 2005 ●6,099 have yet to attempt MCCQE Part II ●8,514 have passed ●718 or 7.7% failed 2,243 candidates were eligible prior to 2000 and also took MCCQE Part II in 2000 – 2005 ●2,166 have passed ●77 or 3.4% failed and are likely out of the system Fail rates do not reflect impact on repeat takers ●Focused hundreds of candidates on remediation

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Limitation Current approach is easy to implement but it relies upon ●Large number of standard setters per case ●Large number of test takers in borderline group Smaller numbers would lead to more effort ●Increase training of examiners ●Impose stricter selection criteria on standard setters

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona What’s ahead? Increasing number of candidates to be assessed each year ●Modifications to the administration are needed Predictive validity study currently in progress ●Use non-physician examiners? Which type of cases, who sets standard? ●Add more administrations? Case development / challenge of piloting content

A Sample of References Burrows, P. J., Bingham, L., & Brailovsky, C. A. (1999). A modified contrasting groups method used for setting the passmark in a small scale standardized patient examination. Advances in Health Sciences Education: Theory and Practice, 4, Cohen, A. S., Kane, M. T., & Crooks, T. J. (1999). A generalized examinee-centered method for setting standards on achievement tests. Applied Measurement in Education, 12, Dauphinee, W. D., Blackmore, D. E., Smee, S. M., Rothman, A. I., & Reznick, R. K. (1997). Using the judgments of physician examiners in setting the standards for a national multi-center high stakes OSCE. Advances in Health Sciences Education: Theory and Practice, 2, Haladyna, T. M. & Hess, R. (2000). An evaluation of conjunctive and compensatory standard-setting strategies for test decisions. Educational Assessment, 6, Hambleton, R. K., Jaeger, R. M., Plake, B. S., & Mills, C. (2000). Setting performance standards on complex educational assessments. Applied Psychological Measurement, 24, Jaeger, R. M. (1995). Setting standards for complex performances: An iterative, judgemental policy-capturing strategy. Educational Measurement: Issues and Practice, 14, Kane, M. T., Crooks, T. J., & Cohen, A. S. (1999). Designing and evaluating standard-setting procedures for licensure and certification tests. Advances in Health Sciences Education: Theory and Practice, 4, Kaufman, D. M., Mann, K. V., Muijtjens, A. M. M., & van der Vleuten, C. P. M. (2001). A comparison of standard-setting procedures for an OSCE in undergraduate medical education. Academic Medicine, 75, Plake, B. S. (1998). Setting performance standards for professional licensure and certification. Applied Measurement in Education, 11, Smee, S. M. & Blackmore, D. E. (2002). Setting standards for an objective structured clinical examination: The borderline group method gains ground on Angoff. Medical Education, 35, Southgate, L., Hays, R. B., Norcini, J. J., Mulholland, H., Ayers, B., Woolliscroft, J., Cusimano, M. D., MacAvoy, P., Ainsworth, M., Haist, S., & Campbell, M. (2001). Setting performance standards for medical practice: A theoretical framework. Medical Education, 35, Wilkinson, T. J., Newble, D. I., & Frampton, C. M. (2001). Standard setting in an objective structured clinical examination: use of global ratings of borderline performance to determine the passing score. Medical Education, 35, Zieky, M. J. (2001). So much has changed: How the setting of cutscores has evolved since the 1980s. In G.J.Cizek (Ed.), Setting performance standards: Concepts, methods, and perspectives (pp ). Mahwah, NJ: Lawrence Erlbaum Associates.

Presented at the 2005 CLEAR Annual Conference September Phoenix, Arizona Medical Council of Canada Ottawa Sydney M. Smee, M.Ed. Manager, MCCQE Part II